@azure/synapse-spark 1.0.0-alpha.20250619.1 → 1.0.0-alpha.20250721.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1 +1 @@
1
- {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAM,OAAO,cAAc;IAGzB;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
1
+ {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAM,OAAO,cAAc;IACR,MAAM,CAAc;IAErC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
@@ -155,6 +155,7 @@ const cancelSparkStatementOperationSpec = {
155
155
  };
156
156
  /** Class containing SparkSessionOperations operations. */
157
157
  export class SparkSessionOperationsImpl {
158
+ client;
158
159
  /**
159
160
  * Initialize a new instance of the class SparkSessionOperations class.
160
161
  * @param client - Reference to the service client
@@ -167,7 +168,7 @@ export class SparkSessionOperationsImpl {
167
168
  * @param options - The options parameters.
168
169
  */
169
170
  async getSparkSessions(options) {
170
- return tracingClient.withSpan("SparkClient.getSparkSessions", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
171
+ return tracingClient.withSpan("SparkClient.getSparkSessions", options ?? {}, async (updatedOptions) => {
171
172
  return this.client.sendOperationRequest({ updatedOptions }, getSparkSessionsOperationSpec);
172
173
  });
173
174
  }
@@ -177,7 +178,7 @@ export class SparkSessionOperationsImpl {
177
178
  * @param options - The options parameters.
178
179
  */
179
180
  async createSparkSession(sparkSessionOptions, options) {
180
- return tracingClient.withSpan("SparkClient.createSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
181
+ return tracingClient.withSpan("SparkClient.createSparkSession", options ?? {}, async (updatedOptions) => {
181
182
  return this.client.sendOperationRequest({ sparkSessionOptions, updatedOptions }, createSparkSessionOperationSpec);
182
183
  });
183
184
  }
@@ -187,7 +188,7 @@ export class SparkSessionOperationsImpl {
187
188
  * @param options - The options parameters.
188
189
  */
189
190
  async getSparkSession(sessionId, options) {
190
- return tracingClient.withSpan("SparkClient.getSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
191
+ return tracingClient.withSpan("SparkClient.getSparkSession", options ?? {}, async (updatedOptions) => {
191
192
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkSessionOperationSpec);
192
193
  });
193
194
  }
@@ -197,7 +198,7 @@ export class SparkSessionOperationsImpl {
197
198
  * @param options - The options parameters.
198
199
  */
199
200
  async cancelSparkSession(sessionId, options) {
200
- return tracingClient.withSpan("SparkClient.cancelSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
201
+ return tracingClient.withSpan("SparkClient.cancelSparkSession", options ?? {}, async (updatedOptions) => {
201
202
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, cancelSparkSessionOperationSpec);
202
203
  });
203
204
  }
@@ -207,7 +208,7 @@ export class SparkSessionOperationsImpl {
207
208
  * @param options - The options parameters.
208
209
  */
209
210
  async resetSparkSessionTimeout(sessionId, options) {
210
- return tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
211
+ return tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options ?? {}, async (updatedOptions) => {
211
212
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, resetSparkSessionTimeoutOperationSpec);
212
213
  });
213
214
  }
@@ -217,7 +218,7 @@ export class SparkSessionOperationsImpl {
217
218
  * @param options - The options parameters.
218
219
  */
219
220
  async getSparkStatements(sessionId, options) {
220
- return tracingClient.withSpan("SparkClient.getSparkStatements", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
221
+ return tracingClient.withSpan("SparkClient.getSparkStatements", options ?? {}, async (updatedOptions) => {
221
222
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkStatementsOperationSpec);
222
223
  });
223
224
  }
@@ -228,7 +229,7 @@ export class SparkSessionOperationsImpl {
228
229
  * @param options - The options parameters.
229
230
  */
230
231
  async createSparkStatement(sessionId, sparkStatementOptions, options) {
231
- return tracingClient.withSpan("SparkClient.createSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
232
+ return tracingClient.withSpan("SparkClient.createSparkStatement", options ?? {}, async (updatedOptions) => {
232
233
  return this.client.sendOperationRequest({ sessionId, sparkStatementOptions, updatedOptions }, createSparkStatementOperationSpec);
233
234
  });
234
235
  }
@@ -239,7 +240,7 @@ export class SparkSessionOperationsImpl {
239
240
  * @param options - The options parameters.
240
241
  */
241
242
  async getSparkStatement(sessionId, statementId, options) {
242
- return tracingClient.withSpan("SparkClient.getSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
243
+ return tracingClient.withSpan("SparkClient.getSparkStatement", options ?? {}, async (updatedOptions) => {
243
244
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, getSparkStatementOperationSpec);
244
245
  });
245
246
  }
@@ -250,7 +251,7 @@ export class SparkSessionOperationsImpl {
250
251
  * @param options - The options parameters.
251
252
  */
252
253
  async cancelSparkStatement(sessionId, statementId, options) {
253
- return tracingClient.withSpan("SparkClient.cancelSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
254
+ return tracingClient.withSpan("SparkClient.cancelSparkStatement", options ?? {}, async (updatedOptions) => {
254
255
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, cancelSparkStatementOperationSpec);
255
256
  });
256
257
  }
@@ -1 +1 @@
1
- {"version":3,"file":"sparkSessionOperations.js","sourceRoot":"","sources":["../../../src/operations/sparkSessionOperations.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAuBtD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,sBAAsB;SAC3C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,mBAAmB;IAC3C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,4BAA4B,GAA6B;IAC7D,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,qCAAqC,GAA6B;IACtE,IAAI,EAAE,kGAAkG;IACxG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,wBAAwB;SAC7C;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,qBAAqB;IAC7C,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,6GAA6G;IACnH,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,oHAAoH;IAC1H,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,gCAAgC;SACrD;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AAEF,0DAA0D;AAC1D,MAAM,OAAO,0BAA0B;IAGrC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAoD;QAEpD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,6BAA6B,CACmB,CAAC;QACrD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,mBAAwC,EACxC,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,mBAAmB,EAAE,cAAc,EAAE,EACvC,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,eAAe,CACnB,SAAiB,EACjB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,6BAA6B,EAC7B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,4BAA4B,CACmB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACf,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,wBAAwB,CAC5B,SAAiB,EACjB,OAA4D;QAE5D,OAAO,aAAa,CAAC,QAAQ,CAC3B,sCAAsC,EACtC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,qCAAqC,CACrB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,qBAA4C,EAC5C,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,qBAAqB,EAAE,cAAc,EAAE,EACpD,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,iBAAiB,CACrB,SAAiB,EACjB,WAAmB,EACnB,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,8BAA8B,CACmB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,WAAmB,EACnB,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkSessionOperations } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkSessionGetSparkSessionsOptionalParams,\n SparkSessionGetSparkSessionsResponse,\n SparkSessionOptions,\n SparkSessionCreateSparkSessionOptionalParams,\n SparkSessionCreateSparkSessionResponse,\n SparkSessionGetSparkSessionOptionalParams,\n SparkSessionGetSparkSessionResponse,\n SparkSessionCancelSparkSessionOptionalParams,\n SparkSessionResetSparkSessionTimeoutOptionalParams,\n SparkSessionGetSparkStatementsOptionalParams,\n SparkSessionGetSparkStatementsResponse,\n SparkStatementOptions,\n SparkSessionCreateSparkStatementOptionalParams,\n SparkSessionCreateSparkStatementResponse,\n SparkSessionGetSparkStatementOptionalParams,\n SparkSessionGetSparkStatementResponse,\n SparkSessionCancelSparkStatementOptionalParams,\n SparkSessionCancelSparkStatementResponse,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkSessionsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSessionCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n requestBody: Parameters.sparkSessionOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst resetSparkSessionTimeoutOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/reset-timeout\",\n httpMethod: \"PUT\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst getSparkStatementsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCollection,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n requestBody: Parameters.sparkStatementOptions,\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}/cancel\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCancellationResult,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\n\n/** Class containing SparkSessionOperations operations. */\nexport class SparkSessionOperationsImpl implements SparkSessionOperations {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkSessionOperations class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark sessions which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkSessions(\n options?: SparkSessionGetSparkSessionsOptionalParams,\n ): Promise<SparkSessionGetSparkSessionsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSessions\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkSessionsOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionsResponse>;\n },\n );\n }\n\n /**\n * Create new spark session.\n * @param sparkSessionOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkSession(\n sparkSessionOptions: SparkSessionOptions,\n options?: SparkSessionCreateSparkSessionOptionalParams,\n ): Promise<SparkSessionCreateSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkSessionOptions, updatedOptions },\n createSparkSessionOperationSpec,\n ) as Promise<SparkSessionCreateSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkSession(\n sessionId: number,\n options?: SparkSessionGetSparkSessionOptionalParams,\n ): Promise<SparkSessionGetSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkSessionOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async cancelSparkSession(\n sessionId: number,\n options?: SparkSessionCancelSparkSessionOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n cancelSparkSessionOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Sends a keep alive call to the current session to reset the session timeout.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async resetSparkSessionTimeout(\n sessionId: number,\n options?: SparkSessionResetSparkSessionTimeoutOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.resetSparkSessionTimeout\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n resetSparkSessionTimeoutOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Gets a list of statements within a spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkStatements(\n sessionId: number,\n options?: SparkSessionGetSparkStatementsOptionalParams,\n ): Promise<SparkSessionGetSparkStatementsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatements\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkStatementsOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementsResponse>;\n },\n );\n }\n\n /**\n * Create statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param sparkStatementOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkStatement(\n sessionId: number,\n sparkStatementOptions: SparkStatementOptions,\n options?: SparkSessionCreateSparkStatementOptionalParams,\n ): Promise<SparkSessionCreateSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, sparkStatementOptions, updatedOptions },\n createSparkStatementOperationSpec,\n ) as Promise<SparkSessionCreateSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Gets a single statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async getSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionGetSparkStatementOptionalParams,\n ): Promise<SparkSessionGetSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n getSparkStatementOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Kill a statement within a session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async cancelSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionCancelSparkStatementOptionalParams,\n ): Promise<SparkSessionCancelSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n cancelSparkStatementOperationSpec,\n ) as Promise<SparkSessionCancelSparkStatementResponse>;\n },\n );\n }\n}\n"]}
1
+ {"version":3,"file":"sparkSessionOperations.js","sourceRoot":"","sources":["../../../src/operations/sparkSessionOperations.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAuBtD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,sBAAsB;SAC3C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,mBAAmB;IAC3C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,4BAA4B,GAA6B;IAC7D,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,qCAAqC,GAA6B;IACtE,IAAI,EAAE,kGAAkG;IACxG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,wBAAwB;SAC7C;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,qBAAqB;IAC7C,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,6GAA6G;IACnH,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,oHAAoH;IAC1H,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,gCAAgC;SACrD;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AAEF,0DAA0D;AAC1D,MAAM,OAAO,0BAA0B;IACpB,MAAM,CAAc;IAErC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAoD;QAEpD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,6BAA6B,CACmB,CAAC;QACrD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,mBAAwC,EACxC,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,mBAAmB,EAAE,cAAc,EAAE,EACvC,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,eAAe,CACnB,SAAiB,EACjB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,6BAA6B,EAC7B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,4BAA4B,CACmB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACf,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,wBAAwB,CAC5B,SAAiB,EACjB,OAA4D;QAE5D,OAAO,aAAa,CAAC,QAAQ,CAC3B,sCAAsC,EACtC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,qCAAqC,CACrB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,qBAA4C,EAC5C,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,qBAAqB,EAAE,cAAc,EAAE,EACpD,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,iBAAiB,CACrB,SAAiB,EACjB,WAAmB,EACnB,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,8BAA8B,CACmB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,WAAmB,EACnB,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkSessionOperations } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkSessionGetSparkSessionsOptionalParams,\n SparkSessionGetSparkSessionsResponse,\n SparkSessionOptions,\n SparkSessionCreateSparkSessionOptionalParams,\n SparkSessionCreateSparkSessionResponse,\n SparkSessionGetSparkSessionOptionalParams,\n SparkSessionGetSparkSessionResponse,\n SparkSessionCancelSparkSessionOptionalParams,\n SparkSessionResetSparkSessionTimeoutOptionalParams,\n SparkSessionGetSparkStatementsOptionalParams,\n SparkSessionGetSparkStatementsResponse,\n SparkStatementOptions,\n SparkSessionCreateSparkStatementOptionalParams,\n SparkSessionCreateSparkStatementResponse,\n SparkSessionGetSparkStatementOptionalParams,\n SparkSessionGetSparkStatementResponse,\n SparkSessionCancelSparkStatementOptionalParams,\n SparkSessionCancelSparkStatementResponse,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkSessionsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSessionCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n requestBody: Parameters.sparkSessionOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst resetSparkSessionTimeoutOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/reset-timeout\",\n httpMethod: \"PUT\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst getSparkStatementsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCollection,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n requestBody: Parameters.sparkStatementOptions,\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}/cancel\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCancellationResult,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\n\n/** Class containing SparkSessionOperations operations. */\nexport class SparkSessionOperationsImpl implements SparkSessionOperations {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkSessionOperations class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark sessions which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkSessions(\n options?: SparkSessionGetSparkSessionsOptionalParams,\n ): Promise<SparkSessionGetSparkSessionsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSessions\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkSessionsOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionsResponse>;\n },\n );\n }\n\n /**\n * Create new spark session.\n * @param sparkSessionOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkSession(\n sparkSessionOptions: SparkSessionOptions,\n options?: SparkSessionCreateSparkSessionOptionalParams,\n ): Promise<SparkSessionCreateSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkSessionOptions, updatedOptions },\n createSparkSessionOperationSpec,\n ) as Promise<SparkSessionCreateSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkSession(\n sessionId: number,\n options?: SparkSessionGetSparkSessionOptionalParams,\n ): Promise<SparkSessionGetSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkSessionOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async cancelSparkSession(\n sessionId: number,\n options?: SparkSessionCancelSparkSessionOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n cancelSparkSessionOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Sends a keep alive call to the current session to reset the session timeout.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async resetSparkSessionTimeout(\n sessionId: number,\n options?: SparkSessionResetSparkSessionTimeoutOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.resetSparkSessionTimeout\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n resetSparkSessionTimeoutOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Gets a list of statements within a spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkStatements(\n sessionId: number,\n options?: SparkSessionGetSparkStatementsOptionalParams,\n ): Promise<SparkSessionGetSparkStatementsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatements\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkStatementsOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementsResponse>;\n },\n );\n }\n\n /**\n * Create statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param sparkStatementOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkStatement(\n sessionId: number,\n sparkStatementOptions: SparkStatementOptions,\n options?: SparkSessionCreateSparkStatementOptionalParams,\n ): Promise<SparkSessionCreateSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, sparkStatementOptions, updatedOptions },\n createSparkStatementOperationSpec,\n ) as Promise<SparkSessionCreateSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Gets a single statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async getSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionGetSparkStatementOptionalParams,\n ): Promise<SparkSessionGetSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n getSparkStatementOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Kill a statement within a session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async cancelSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionCancelSparkStatementOptionalParams,\n ): Promise<SparkSessionCancelSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n cancelSparkStatementOperationSpec,\n ) as Promise<SparkSessionCancelSparkStatementResponse>;\n },\n );\n }\n}\n"]}
@@ -12,6 +12,9 @@ import { SparkBatchImpl, SparkSessionOperationsImpl } from "./operations/index.j
12
12
  * Represents the Synapse Spark client operations.
13
13
  */
14
14
  export class SparkClient extends coreClient.ServiceClient {
15
+ endpoint;
16
+ livyApiVersion;
17
+ sparkPoolName;
15
18
  /**
16
19
  * Initializes a new instance of the SparkClient class.
17
20
  * @param credentials - Subscription credentials which uniquely identify client subscription.
@@ -21,7 +24,6 @@ export class SparkClient extends coreClient.ServiceClient {
21
24
  * @param options - The parameter options
22
25
  */
23
26
  constructor(credentials, endpoint, sparkPoolName, options) {
24
- var _a, _b;
25
27
  if (credentials === undefined) {
26
28
  throw new Error("'credentials' cannot be null");
27
29
  }
@@ -46,11 +48,16 @@ export class SparkClient extends coreClient.ServiceClient {
46
48
  if (!options.credentialScopes) {
47
49
  options.credentialScopes = ["https://dev.azuresynapse.net/.default"];
48
50
  }
49
- const optionsWithDefaults = Object.assign(Object.assign(Object.assign({}, defaults), options), { userAgentOptions: {
51
+ const optionsWithDefaults = {
52
+ ...defaults,
53
+ ...options,
54
+ userAgentOptions: {
50
55
  userAgentPrefix,
51
- }, baseUri: (_b = (_a = options.endpoint) !== null && _a !== void 0 ? _a : options.baseUri) !== null && _b !== void 0 ? _b : "{endpoint}" });
56
+ },
57
+ baseUri: options.endpoint ?? options.baseUri ?? "{endpoint}",
58
+ };
52
59
  super(optionsWithDefaults);
53
- if ((options === null || options === void 0 ? void 0 : options.pipeline) && options.pipeline.getOrderedPolicies().length > 0) {
60
+ if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {
54
61
  const pipelinePolicies = options.pipeline.getOrderedPolicies();
55
62
  const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some((pipelinePolicy) => pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName);
56
63
  if (!bearerTokenAuthenticationPolicyFound) {
@@ -73,5 +80,7 @@ export class SparkClient extends coreClient.ServiceClient {
73
80
  this.sparkBatch = new SparkBatchImpl(this);
74
81
  this.sparkSessionOperations = new SparkSessionOperationsImpl(this);
75
82
  }
83
+ sparkBatch;
84
+ sparkSessionOperations;
76
85
  }
77
86
  //# sourceMappingURL=sparkClient.js.map
@@ -1 +1 @@
1
- {"version":3,"file":"sparkClient.js","sourceRoot":"","sources":["../../src/sparkClient.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,gBAAgB,MAAM,2BAA2B,CAAC;AAE9D,OAAO,EAAE,cAAc,EAAE,0BAA0B,EAAE,MAAM,uBAAuB,CAAC;AAInF;;GAEG;AACH,MAAM,OAAO,WAAY,SAAQ,UAAU,CAAC,aAAa;IAKvD;;;;;;;OAOG;IACH,YACE,WAAqC,EACrC,QAAgB,EAChB,aAAqB,EACrB,OAAmC;;QAEnC,IAAI,WAAW,KAAK,SAAS,EAAE,CAAC;YAC9B,MAAM,IAAI,KAAK,CAAC,8BAA8B,CAAC,CAAC;QAClD,CAAC;QACD,IAAI,QAAQ,KAAK,SAAS,EAAE,CAAC;YAC3B,MAAM,IAAI,KAAK,CAAC,2BAA2B,CAAC,CAAC;QAC/C,CAAC;QACD,IAAI,aAAa,KAAK,SAAS,EAAE,CAAC;YAChC,MAAM,IAAI,KAAK,CAAC,gCAAgC,CAAC,CAAC;QACpD,CAAC;QAED,0CAA0C;QAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;YACb,OAAO,GAAG,EAAE,CAAC;QACf,CAAC;QACD,MAAM,QAAQ,GAA8B;YAC1C,kBAAkB,EAAE,iCAAiC;YACrD,UAAU,EAAE,WAAW;SACxB,CAAC;QAEF,MAAM,cAAc,GAAG,qCAAqC,CAAC;QAC7D,MAAM,eAAe,GACnB,OAAO,CAAC,gBAAgB,IAAI,OAAO,CAAC,gBAAgB,CAAC,eAAe;YAClE,CAAC,CAAC,GAAG,OAAO,CAAC,gBAAgB,CAAC,eAAe,IAAI,cAAc,EAAE;YACjE,CAAC,CAAC,GAAG,cAAc,EAAE,CAAC;QAE1B,IAAI,CAAC,OAAO,CAAC,gBAAgB,EAAE,CAAC;YAC9B,OAAO,CAAC,gBAAgB,GAAG,CAAC,uCAAuC,CAAC,CAAC;QACvE,CAAC;QACD,MAAM,mBAAmB,iDACpB,QAAQ,GACR,OAAO,KACV,gBAAgB,EAAE;gBAChB,eAAe;aAChB,EACD,OAAO,EAAE,MAAA,MAAA,OAAO,CAAC,QAAQ,mCAAI,OAAO,CAAC,OAAO,mCAAI,YAAY,GAC7D,CAAC;QACF,KAAK,CAAC,mBAAmB,CAAC,CAAC;QAE3B,IAAI,CAAA,OAAO,aAAP,OAAO,uBAAP,OAAO,CAAE,QAAQ,KAAI,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YAC1E,MAAM,gBAAgB,GACpB,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC;YACxC,MAAM,oCAAoC,GAAG,gBAAgB,CAAC,IAAI,CAChE,CAAC,cAAc,EAAE,EAAE,CACjB,cAAc,CAAC,IAAI,KAAK,gBAAgB,CAAC,mCAAmC,CAC/E,CAAC;YACF,IAAI,CAAC,oCAAoC,EAAE,CAAC;gBAC1C,IAAI,CAAC,QAAQ,CAAC,YAAY,CAAC;oBACzB,IAAI,EAAE,gBAAgB,CAAC,mCAAmC;iBAC3D,CAAC,CAAC;gBACH,IAAI,CAAC,QAAQ,CAAC,SAAS,CACrB,gBAAgB,CAAC,+BAA+B,CAAC;oBAC/C,MAAM,EAAE,GAAG,mBAAmB,CAAC,OAAO,WAAW;oBACjD,kBAAkB,EAAE;wBAClB,2BAA2B,EAAE,UAAU,CAAC,gCAAgC;qBACzE;iBACF,CAAC,CACH,CAAC;YACJ,CAAC;QACH,CAAC;QACD,wBAAwB;QACxB,IAAI,CAAC,QAAQ,GAAG,QAAQ,CAAC;QACzB,IAAI,CAAC,aAAa,GAAG,aAAa,CAAC;QAEnC,0CAA0C;QAC1C,IAAI,CAAC,cAAc,GAAG,OAAO,CAAC,cAAc,IAAI,oBAAoB,CAAC;QACrE,IAAI,CAAC,UAAU,GAAG,IAAI,cAAc,CAAC,IAAI,CAAC,CAAC;QAC3C,IAAI,CAAC,sBAAsB,GAAG,IAAI,0BAA0B,CAAC,IAAI,CAAC,CAAC;IACrE,CAAC;CAIF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport * as coreClient from \"@azure/core-client\";\nimport * as coreRestPipeline from \"@azure/core-rest-pipeline\";\nimport type * as coreAuth from \"@azure/core-auth\";\nimport { SparkBatchImpl, SparkSessionOperationsImpl } from \"./operations/index.js\";\nimport type { SparkBatch, SparkSessionOperations } from \"./operationsInterfaces/index.js\";\nimport type { SparkClientOptionalParams } from \"./models/index.js\";\n\n/**\n * Represents the Synapse Spark client operations.\n */\nexport class SparkClient extends coreClient.ServiceClient {\n endpoint: string;\n livyApiVersion: string;\n sparkPoolName: string;\n\n /**\n * Initializes a new instance of the SparkClient class.\n * @param credentials - Subscription credentials which uniquely identify client subscription.\n * @param endpoint - The workspace development endpoint, for example\n * https://myworkspace.dev.azuresynapse.net.\n * @param sparkPoolName - Name of the spark pool.\n * @param options - The parameter options\n */\n constructor(\n credentials: coreAuth.TokenCredential,\n endpoint: string,\n sparkPoolName: string,\n options?: SparkClientOptionalParams,\n ) {\n if (credentials === undefined) {\n throw new Error(\"'credentials' cannot be null\");\n }\n if (endpoint === undefined) {\n throw new Error(\"'endpoint' cannot be null\");\n }\n if (sparkPoolName === undefined) {\n throw new Error(\"'sparkPoolName' cannot be null\");\n }\n\n // Initializing default values for options\n if (!options) {\n options = {};\n }\n const defaults: SparkClientOptionalParams = {\n requestContentType: \"application/json; charset=utf-8\",\n credential: credentials,\n };\n\n const packageDetails = `azsdk-js-synapse-spark/1.0.0-beta.6`;\n const userAgentPrefix =\n options.userAgentOptions && options.userAgentOptions.userAgentPrefix\n ? `${options.userAgentOptions.userAgentPrefix} ${packageDetails}`\n : `${packageDetails}`;\n\n if (!options.credentialScopes) {\n options.credentialScopes = [\"https://dev.azuresynapse.net/.default\"];\n }\n const optionsWithDefaults = {\n ...defaults,\n ...options,\n userAgentOptions: {\n userAgentPrefix,\n },\n baseUri: options.endpoint ?? options.baseUri ?? \"{endpoint}\",\n };\n super(optionsWithDefaults);\n\n if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {\n const pipelinePolicies: coreRestPipeline.PipelinePolicy[] =\n options.pipeline.getOrderedPolicies();\n const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some(\n (pipelinePolicy) =>\n pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName,\n );\n if (!bearerTokenAuthenticationPolicyFound) {\n this.pipeline.removePolicy({\n name: coreRestPipeline.bearerTokenAuthenticationPolicyName,\n });\n this.pipeline.addPolicy(\n coreRestPipeline.bearerTokenAuthenticationPolicy({\n scopes: `${optionsWithDefaults.baseUri}/.default`,\n challengeCallbacks: {\n authorizeRequestOnChallenge: coreClient.authorizeRequestOnClaimChallenge,\n },\n }),\n );\n }\n }\n // Parameter assignments\n this.endpoint = endpoint;\n this.sparkPoolName = sparkPoolName;\n\n // Assigning values to Constant parameters\n this.livyApiVersion = options.livyApiVersion || \"2019-11-01-preview\";\n this.sparkBatch = new SparkBatchImpl(this);\n this.sparkSessionOperations = new SparkSessionOperationsImpl(this);\n }\n\n sparkBatch: SparkBatch;\n sparkSessionOperations: SparkSessionOperations;\n}\n"]}
1
+ {"version":3,"file":"sparkClient.js","sourceRoot":"","sources":["../../src/sparkClient.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,gBAAgB,MAAM,2BAA2B,CAAC;AAE9D,OAAO,EAAE,cAAc,EAAE,0BAA0B,EAAE,MAAM,uBAAuB,CAAC;AAInF;;GAEG;AACH,MAAM,OAAO,WAAY,SAAQ,UAAU,CAAC,aAAa;IACvD,QAAQ,CAAS;IACjB,cAAc,CAAS;IACvB,aAAa,CAAS;IAEtB;;;;;;;OAOG;IACH,YACE,WAAqC,EACrC,QAAgB,EAChB,aAAqB,EACrB,OAAmC;QAEnC,IAAI,WAAW,KAAK,SAAS,EAAE,CAAC;YAC9B,MAAM,IAAI,KAAK,CAAC,8BAA8B,CAAC,CAAC;QAClD,CAAC;QACD,IAAI,QAAQ,KAAK,SAAS,EAAE,CAAC;YAC3B,MAAM,IAAI,KAAK,CAAC,2BAA2B,CAAC,CAAC;QAC/C,CAAC;QACD,IAAI,aAAa,KAAK,SAAS,EAAE,CAAC;YAChC,MAAM,IAAI,KAAK,CAAC,gCAAgC,CAAC,CAAC;QACpD,CAAC;QAED,0CAA0C;QAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;YACb,OAAO,GAAG,EAAE,CAAC;QACf,CAAC;QACD,MAAM,QAAQ,GAA8B;YAC1C,kBAAkB,EAAE,iCAAiC;YACrD,UAAU,EAAE,WAAW;SACxB,CAAC;QAEF,MAAM,cAAc,GAAG,qCAAqC,CAAC;QAC7D,MAAM,eAAe,GACnB,OAAO,CAAC,gBAAgB,IAAI,OAAO,CAAC,gBAAgB,CAAC,eAAe;YAClE,CAAC,CAAC,GAAG,OAAO,CAAC,gBAAgB,CAAC,eAAe,IAAI,cAAc,EAAE;YACjE,CAAC,CAAC,GAAG,cAAc,EAAE,CAAC;QAE1B,IAAI,CAAC,OAAO,CAAC,gBAAgB,EAAE,CAAC;YAC9B,OAAO,CAAC,gBAAgB,GAAG,CAAC,uCAAuC,CAAC,CAAC;QACvE,CAAC;QACD,MAAM,mBAAmB,GAAG;YAC1B,GAAG,QAAQ;YACX,GAAG,OAAO;YACV,gBAAgB,EAAE;gBAChB,eAAe;aAChB;YACD,OAAO,EAAE,OAAO,CAAC,QAAQ,IAAI,OAAO,CAAC,OAAO,IAAI,YAAY;SAC7D,CAAC;QACF,KAAK,CAAC,mBAAmB,CAAC,CAAC;QAE3B,IAAI,OAAO,EAAE,QAAQ,IAAI,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YAC1E,MAAM,gBAAgB,GACpB,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC;YACxC,MAAM,oCAAoC,GAAG,gBAAgB,CAAC,IAAI,CAChE,CAAC,cAAc,EAAE,EAAE,CACjB,cAAc,CAAC,IAAI,KAAK,gBAAgB,CAAC,mCAAmC,CAC/E,CAAC;YACF,IAAI,CAAC,oCAAoC,EAAE,CAAC;gBAC1C,IAAI,CAAC,QAAQ,CAAC,YAAY,CAAC;oBACzB,IAAI,EAAE,gBAAgB,CAAC,mCAAmC;iBAC3D,CAAC,CAAC;gBACH,IAAI,CAAC,QAAQ,CAAC,SAAS,CACrB,gBAAgB,CAAC,+BAA+B,CAAC;oBAC/C,MAAM,EAAE,GAAG,mBAAmB,CAAC,OAAO,WAAW;oBACjD,kBAAkB,EAAE;wBAClB,2BAA2B,EAAE,UAAU,CAAC,gCAAgC;qBACzE;iBACF,CAAC,CACH,CAAC;YACJ,CAAC;QACH,CAAC;QACD,wBAAwB;QACxB,IAAI,CAAC,QAAQ,GAAG,QAAQ,CAAC;QACzB,IAAI,CAAC,aAAa,GAAG,aAAa,CAAC;QAEnC,0CAA0C;QAC1C,IAAI,CAAC,cAAc,GAAG,OAAO,CAAC,cAAc,IAAI,oBAAoB,CAAC;QACrE,IAAI,CAAC,UAAU,GAAG,IAAI,cAAc,CAAC,IAAI,CAAC,CAAC;QAC3C,IAAI,CAAC,sBAAsB,GAAG,IAAI,0BAA0B,CAAC,IAAI,CAAC,CAAC;IACrE,CAAC;IAED,UAAU,CAAa;IACvB,sBAAsB,CAAyB;CAChD","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport * as coreClient from \"@azure/core-client\";\nimport * as coreRestPipeline from \"@azure/core-rest-pipeline\";\nimport type * as coreAuth from \"@azure/core-auth\";\nimport { SparkBatchImpl, SparkSessionOperationsImpl } from \"./operations/index.js\";\nimport type { SparkBatch, SparkSessionOperations } from \"./operationsInterfaces/index.js\";\nimport type { SparkClientOptionalParams } from \"./models/index.js\";\n\n/**\n * Represents the Synapse Spark client operations.\n */\nexport class SparkClient extends coreClient.ServiceClient {\n endpoint: string;\n livyApiVersion: string;\n sparkPoolName: string;\n\n /**\n * Initializes a new instance of the SparkClient class.\n * @param credentials - Subscription credentials which uniquely identify client subscription.\n * @param endpoint - The workspace development endpoint, for example\n * https://myworkspace.dev.azuresynapse.net.\n * @param sparkPoolName - Name of the spark pool.\n * @param options - The parameter options\n */\n constructor(\n credentials: coreAuth.TokenCredential,\n endpoint: string,\n sparkPoolName: string,\n options?: SparkClientOptionalParams,\n ) {\n if (credentials === undefined) {\n throw new Error(\"'credentials' cannot be null\");\n }\n if (endpoint === undefined) {\n throw new Error(\"'endpoint' cannot be null\");\n }\n if (sparkPoolName === undefined) {\n throw new Error(\"'sparkPoolName' cannot be null\");\n }\n\n // Initializing default values for options\n if (!options) {\n options = {};\n }\n const defaults: SparkClientOptionalParams = {\n requestContentType: \"application/json; charset=utf-8\",\n credential: credentials,\n };\n\n const packageDetails = `azsdk-js-synapse-spark/1.0.0-beta.6`;\n const userAgentPrefix =\n options.userAgentOptions && options.userAgentOptions.userAgentPrefix\n ? `${options.userAgentOptions.userAgentPrefix} ${packageDetails}`\n : `${packageDetails}`;\n\n if (!options.credentialScopes) {\n options.credentialScopes = [\"https://dev.azuresynapse.net/.default\"];\n }\n const optionsWithDefaults = {\n ...defaults,\n ...options,\n userAgentOptions: {\n userAgentPrefix,\n },\n baseUri: options.endpoint ?? options.baseUri ?? \"{endpoint}\",\n };\n super(optionsWithDefaults);\n\n if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {\n const pipelinePolicies: coreRestPipeline.PipelinePolicy[] =\n options.pipeline.getOrderedPolicies();\n const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some(\n (pipelinePolicy) =>\n pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName,\n );\n if (!bearerTokenAuthenticationPolicyFound) {\n this.pipeline.removePolicy({\n name: coreRestPipeline.bearerTokenAuthenticationPolicyName,\n });\n this.pipeline.addPolicy(\n coreRestPipeline.bearerTokenAuthenticationPolicy({\n scopes: `${optionsWithDefaults.baseUri}/.default`,\n challengeCallbacks: {\n authorizeRequestOnChallenge: coreClient.authorizeRequestOnClaimChallenge,\n },\n }),\n );\n }\n }\n // Parameter assignments\n this.endpoint = endpoint;\n this.sparkPoolName = sparkPoolName;\n\n // Assigning values to Constant parameters\n this.livyApiVersion = options.livyApiVersion || \"2019-11-01-preview\";\n this.sparkBatch = new SparkBatchImpl(this);\n this.sparkSessionOperations = new SparkSessionOperationsImpl(this);\n }\n\n sparkBatch: SparkBatch;\n sparkSessionOperations: SparkSessionOperations;\n}\n"]}
package/package.json CHANGED
@@ -5,7 +5,7 @@
5
5
  "homepage": "https://github.com/Azure/azure-sdk-for-js/blob/main/sdk/synapse/synapse-spark/README.md",
6
6
  "repository": "github:Azure/azure-sdk-for-js",
7
7
  "sdk-type": "client",
8
- "version": "1.0.0-alpha.20250619.1",
8
+ "version": "1.0.0-alpha.20250721.1",
9
9
  "dependencies": {
10
10
  "@azure/core-auth": "^1.9.0",
11
11
  "@azure/core-client": "^1.9.2",
@@ -85,7 +85,7 @@
85
85
  "autoPublish": true,
86
86
  "type": "module",
87
87
  "tshy": {
88
- "project": "./tsconfig.src.json",
88
+ "project": "../../../tsconfig.src.build.json",
89
89
  "exports": {
90
90
  "./package.json": "./package.json",
91
91
  ".": "./src/index.ts"