@wix/auto_sdk_seo_robots-txt 1.0.17 → 1.0.19

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,32 @@
1
+ import * as z from 'zod';
2
+
3
+ declare const GetRobotsTxtRequest: z.ZodObject<{
4
+ options: z.ZodOptional<z.ZodObject<{
5
+ subdomain: z.ZodOptional<z.ZodString>;
6
+ }, z.core.$strip>>;
7
+ }, z.core.$strip>;
8
+ declare const GetRobotsTxtResponse: z.ZodObject<{
9
+ robotsTxt: z.ZodOptional<z.ZodObject<{
10
+ content: z.ZodOptional<z.ZodString>;
11
+ default: z.ZodOptional<z.ZodBoolean>;
12
+ subdomain: z.ZodOptional<z.ZodString>;
13
+ }, z.core.$strip>>;
14
+ }, z.core.$strip>;
15
+ declare const UpdateRobotsTxtRequest: z.ZodObject<{
16
+ options: z.ZodOptional<z.ZodObject<{
17
+ robotsTxt: z.ZodOptional<z.ZodObject<{
18
+ content: z.ZodOptional<z.ZodString>;
19
+ default: z.ZodOptional<z.ZodBoolean>;
20
+ subdomain: z.ZodOptional<z.ZodString>;
21
+ }, z.core.$strip>>;
22
+ }, z.core.$strip>>;
23
+ }, z.core.$strip>;
24
+ declare const UpdateRobotsTxtResponse: z.ZodObject<{
25
+ robotsTxt: z.ZodOptional<z.ZodObject<{
26
+ content: z.ZodOptional<z.ZodString>;
27
+ default: z.ZodOptional<z.ZodBoolean>;
28
+ subdomain: z.ZodOptional<z.ZodString>;
29
+ }, z.core.$strip>>;
30
+ }, z.core.$strip>;
31
+
32
+ export { GetRobotsTxtRequest, GetRobotsTxtResponse, UpdateRobotsTxtRequest, UpdateRobotsTxtResponse };
@@ -0,0 +1,91 @@
1
+ "use strict";
2
+ var __create = Object.create;
3
+ var __defProp = Object.defineProperty;
4
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
+ var __getOwnPropNames = Object.getOwnPropertyNames;
6
+ var __getProtoOf = Object.getPrototypeOf;
7
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
8
+ var __export = (target, all) => {
9
+ for (var name in all)
10
+ __defProp(target, name, { get: all[name], enumerable: true });
11
+ };
12
+ var __copyProps = (to, from, except, desc) => {
13
+ if (from && typeof from === "object" || typeof from === "function") {
14
+ for (let key of __getOwnPropNames(from))
15
+ if (!__hasOwnProp.call(to, key) && key !== except)
16
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
17
+ }
18
+ return to;
19
+ };
20
+ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
+ // If the importer is in node compatibility mode or this is not an ESM
22
+ // file that has been converted to a CommonJS file using a Babel-
23
+ // compatible transform (i.e. "__esModule" has not been set), then set
24
+ // "default" to the CommonJS "module.exports" for node compatibility.
25
+ isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
+ mod
27
+ ));
28
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
+
30
+ // schemas.ts
31
+ var schemas_exports = {};
32
+ __export(schemas_exports, {
33
+ GetRobotsTxtRequest: () => GetRobotsTxtRequest,
34
+ GetRobotsTxtResponse: () => GetRobotsTxtResponse,
35
+ UpdateRobotsTxtRequest: () => UpdateRobotsTxtRequest,
36
+ UpdateRobotsTxtResponse: () => UpdateRobotsTxtResponse
37
+ });
38
+ module.exports = __toCommonJS(schemas_exports);
39
+
40
+ // src/promote-seo-v2-robots-txt-robots-txt.schemas.ts
41
+ var z = __toESM(require("zod"));
42
+ var GetRobotsTxtRequest = z.object({
43
+ options: z.object({
44
+ subdomain: z.string().describe(
45
+ "Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`."
46
+ ).optional()
47
+ }).optional()
48
+ });
49
+ var GetRobotsTxtResponse = z.object({
50
+ robotsTxt: z.object({
51
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
52
+ default: z.boolean().describe(
53
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
54
+ ).optional(),
55
+ subdomain: z.string().describe(
56
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
57
+ ).max(63).optional()
58
+ }).describe("Retrieved `robots.txt` file entity.").optional()
59
+ });
60
+ var UpdateRobotsTxtRequest = z.object({
61
+ options: z.object({
62
+ robotsTxt: z.object({
63
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
64
+ default: z.boolean().describe(
65
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
66
+ ).optional(),
67
+ subdomain: z.string().describe(
68
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
69
+ ).max(63).optional()
70
+ }).describe("`robots.txt` file to update.").optional()
71
+ }).optional()
72
+ });
73
+ var UpdateRobotsTxtResponse = z.object({
74
+ robotsTxt: z.object({
75
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
76
+ default: z.boolean().describe(
77
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
78
+ ).optional(),
79
+ subdomain: z.string().describe(
80
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
81
+ ).max(63).optional()
82
+ }).describe("Updated `robots.txt` file.").optional()
83
+ });
84
+ // Annotate the CommonJS export names for ESM import in node:
85
+ 0 && (module.exports = {
86
+ GetRobotsTxtRequest,
87
+ GetRobotsTxtResponse,
88
+ UpdateRobotsTxtRequest,
89
+ UpdateRobotsTxtResponse
90
+ });
91
+ //# sourceMappingURL=schemas.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../schemas.ts","../../src/promote-seo-v2-robots-txt-robots-txt.schemas.ts"],"sourcesContent":["export * from './src/promote-seo-v2-robots-txt-robots-txt.schemas.js';\n","import * as z from 'zod';\n\nexport const GetRobotsTxtRequest = z.object({\n options: z\n .object({\n subdomain: z\n .string()\n .describe(\n 'Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`.'\n )\n .optional(),\n })\n .optional(),\n});\nexport const GetRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Retrieved `robots.txt` file entity.')\n .optional(),\n});\nexport const UpdateRobotsTxtRequest = z.object({\n options: z\n .object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('`robots.txt` file to update.')\n .optional(),\n })\n .optional(),\n});\nexport const UpdateRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Updated `robots.txt` file.')\n .optional(),\n});\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,QAAmB;AAEZ,IAAM,sBAAwB,SAAO;AAAA,EAC1C,SACG,SAAO;AAAA,IACN,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,uBAAyB,SAAO;AAAA,EAC3C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,qCAAqC,EAC9C,SAAS;AACd,CAAC;AACM,IAAM,yBAA2B,SAAO;AAAA,EAC7C,SACG,SAAO;AAAA,IACN,WACG,SAAO;AAAA,MACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,MACZ,SACG,UAAQ,EACR;AAAA,QACC;AAAA,MACF,EACC,SAAS;AAAA,MACZ,WACG,SAAO,EACP;AAAA,QACC;AAAA,MACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,IACd,CAAC,EACA,SAAS,8BAA8B,EACvC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,0BAA4B,SAAO;AAAA,EAC9C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,4BAA4B,EACrC,SAAS;AACd,CAAC;","names":[]}
@@ -0,0 +1,32 @@
1
+ import * as z from 'zod';
2
+
3
+ declare const GetRobotsTxtRequest: z.ZodObject<{
4
+ options: z.ZodOptional<z.ZodObject<{
5
+ subdomain: z.ZodOptional<z.ZodString>;
6
+ }, z.core.$strip>>;
7
+ }, z.core.$strip>;
8
+ declare const GetRobotsTxtResponse: z.ZodObject<{
9
+ robotsTxt: z.ZodOptional<z.ZodObject<{
10
+ content: z.ZodOptional<z.ZodString>;
11
+ default: z.ZodOptional<z.ZodBoolean>;
12
+ subdomain: z.ZodOptional<z.ZodString>;
13
+ }, z.core.$strip>>;
14
+ }, z.core.$strip>;
15
+ declare const UpdateRobotsTxtRequest: z.ZodObject<{
16
+ options: z.ZodOptional<z.ZodObject<{
17
+ robotsTxt: z.ZodOptional<z.ZodObject<{
18
+ content: z.ZodOptional<z.ZodString>;
19
+ default: z.ZodOptional<z.ZodBoolean>;
20
+ subdomain: z.ZodOptional<z.ZodString>;
21
+ }, z.core.$strip>>;
22
+ }, z.core.$strip>>;
23
+ }, z.core.$strip>;
24
+ declare const UpdateRobotsTxtResponse: z.ZodObject<{
25
+ robotsTxt: z.ZodOptional<z.ZodObject<{
26
+ content: z.ZodOptional<z.ZodString>;
27
+ default: z.ZodOptional<z.ZodBoolean>;
28
+ subdomain: z.ZodOptional<z.ZodString>;
29
+ }, z.core.$strip>>;
30
+ }, z.core.$strip>;
31
+
32
+ export { GetRobotsTxtRequest, GetRobotsTxtResponse, UpdateRobotsTxtRequest, UpdateRobotsTxtResponse };
@@ -0,0 +1,51 @@
1
+ // src/promote-seo-v2-robots-txt-robots-txt.schemas.ts
2
+ import * as z from "zod";
3
+ var GetRobotsTxtRequest = z.object({
4
+ options: z.object({
5
+ subdomain: z.string().describe(
6
+ "Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`."
7
+ ).optional()
8
+ }).optional()
9
+ });
10
+ var GetRobotsTxtResponse = z.object({
11
+ robotsTxt: z.object({
12
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
13
+ default: z.boolean().describe(
14
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
15
+ ).optional(),
16
+ subdomain: z.string().describe(
17
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
18
+ ).max(63).optional()
19
+ }).describe("Retrieved `robots.txt` file entity.").optional()
20
+ });
21
+ var UpdateRobotsTxtRequest = z.object({
22
+ options: z.object({
23
+ robotsTxt: z.object({
24
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
25
+ default: z.boolean().describe(
26
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
27
+ ).optional(),
28
+ subdomain: z.string().describe(
29
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
30
+ ).max(63).optional()
31
+ }).describe("`robots.txt` file to update.").optional()
32
+ }).optional()
33
+ });
34
+ var UpdateRobotsTxtResponse = z.object({
35
+ robotsTxt: z.object({
36
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
37
+ default: z.boolean().describe(
38
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
39
+ ).optional(),
40
+ subdomain: z.string().describe(
41
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
42
+ ).max(63).optional()
43
+ }).describe("Updated `robots.txt` file.").optional()
44
+ });
45
+ export {
46
+ GetRobotsTxtRequest,
47
+ GetRobotsTxtResponse,
48
+ UpdateRobotsTxtRequest,
49
+ UpdateRobotsTxtResponse
50
+ };
51
+ //# sourceMappingURL=schemas.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../src/promote-seo-v2-robots-txt-robots-txt.schemas.ts"],"sourcesContent":["import * as z from 'zod';\n\nexport const GetRobotsTxtRequest = z.object({\n options: z\n .object({\n subdomain: z\n .string()\n .describe(\n 'Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`.'\n )\n .optional(),\n })\n .optional(),\n});\nexport const GetRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Retrieved `robots.txt` file entity.')\n .optional(),\n});\nexport const UpdateRobotsTxtRequest = z.object({\n options: z\n .object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('`robots.txt` file to update.')\n .optional(),\n })\n .optional(),\n});\nexport const UpdateRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Updated `robots.txt` file.')\n .optional(),\n});\n"],"mappings":";AAAA,YAAY,OAAO;AAEZ,IAAM,sBAAwB,SAAO;AAAA,EAC1C,SACG,SAAO;AAAA,IACN,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,uBAAyB,SAAO;AAAA,EAC3C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,qCAAqC,EAC9C,SAAS;AACd,CAAC;AACM,IAAM,yBAA2B,SAAO;AAAA,EAC7C,SACG,SAAO;AAAA,IACN,WACG,SAAO;AAAA,MACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,MACZ,SACG,UAAQ,EACR;AAAA,QACC;AAAA,MACF,EACC,SAAS;AAAA,MACZ,WACG,SAAO,EACP;AAAA,QACC;AAAA,MACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,IACd,CAAC,EACA,SAAS,8BAA8B,EACvC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,0BAA4B,SAAO;AAAA,EAC9C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,4BAA4B,EACrC,SAAS;AACd,CAAC;","names":[]}
@@ -0,0 +1,32 @@
1
+ import * as z from 'zod';
2
+
3
+ declare const GetRobotsTxtRequest: z.ZodObject<{
4
+ options: z.ZodOptional<z.ZodObject<{
5
+ subdomain: z.ZodOptional<z.ZodString>;
6
+ }, z.core.$strip>>;
7
+ }, z.core.$strip>;
8
+ declare const GetRobotsTxtResponse: z.ZodObject<{
9
+ robotsTxt: z.ZodOptional<z.ZodObject<{
10
+ content: z.ZodOptional<z.ZodString>;
11
+ default: z.ZodOptional<z.ZodBoolean>;
12
+ subdomain: z.ZodOptional<z.ZodString>;
13
+ }, z.core.$strip>>;
14
+ }, z.core.$strip>;
15
+ declare const UpdateRobotsTxtRequest: z.ZodObject<{
16
+ options: z.ZodOptional<z.ZodObject<{
17
+ robotsTxt: z.ZodOptional<z.ZodObject<{
18
+ content: z.ZodOptional<z.ZodString>;
19
+ default: z.ZodOptional<z.ZodBoolean>;
20
+ subdomain: z.ZodOptional<z.ZodString>;
21
+ }, z.core.$strip>>;
22
+ }, z.core.$strip>>;
23
+ }, z.core.$strip>;
24
+ declare const UpdateRobotsTxtResponse: z.ZodObject<{
25
+ robotsTxt: z.ZodOptional<z.ZodObject<{
26
+ content: z.ZodOptional<z.ZodString>;
27
+ default: z.ZodOptional<z.ZodBoolean>;
28
+ subdomain: z.ZodOptional<z.ZodString>;
29
+ }, z.core.$strip>>;
30
+ }, z.core.$strip>;
31
+
32
+ export { GetRobotsTxtRequest, GetRobotsTxtResponse, UpdateRobotsTxtRequest, UpdateRobotsTxtResponse };
@@ -0,0 +1,91 @@
1
+ "use strict";
2
+ var __create = Object.create;
3
+ var __defProp = Object.defineProperty;
4
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
5
+ var __getOwnPropNames = Object.getOwnPropertyNames;
6
+ var __getProtoOf = Object.getPrototypeOf;
7
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
8
+ var __export = (target, all) => {
9
+ for (var name in all)
10
+ __defProp(target, name, { get: all[name], enumerable: true });
11
+ };
12
+ var __copyProps = (to, from, except, desc) => {
13
+ if (from && typeof from === "object" || typeof from === "function") {
14
+ for (let key of __getOwnPropNames(from))
15
+ if (!__hasOwnProp.call(to, key) && key !== except)
16
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
17
+ }
18
+ return to;
19
+ };
20
+ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
21
+ // If the importer is in node compatibility mode or this is not an ESM
22
+ // file that has been converted to a CommonJS file using a Babel-
23
+ // compatible transform (i.e. "__esModule" has not been set), then set
24
+ // "default" to the CommonJS "module.exports" for node compatibility.
25
+ isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
26
+ mod
27
+ ));
28
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
29
+
30
+ // schemas.ts
31
+ var schemas_exports = {};
32
+ __export(schemas_exports, {
33
+ GetRobotsTxtRequest: () => GetRobotsTxtRequest,
34
+ GetRobotsTxtResponse: () => GetRobotsTxtResponse,
35
+ UpdateRobotsTxtRequest: () => UpdateRobotsTxtRequest,
36
+ UpdateRobotsTxtResponse: () => UpdateRobotsTxtResponse
37
+ });
38
+ module.exports = __toCommonJS(schemas_exports);
39
+
40
+ // src/promote-seo-v2-robots-txt-robots-txt.schemas.ts
41
+ var z = __toESM(require("zod"));
42
+ var GetRobotsTxtRequest = z.object({
43
+ options: z.object({
44
+ subdomain: z.string().describe(
45
+ "Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`."
46
+ ).optional()
47
+ }).optional()
48
+ });
49
+ var GetRobotsTxtResponse = z.object({
50
+ robotsTxt: z.object({
51
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
52
+ default: z.boolean().describe(
53
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
54
+ ).optional(),
55
+ subdomain: z.string().describe(
56
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
57
+ ).max(63).optional()
58
+ }).describe("Retrieved `robots.txt` file entity.").optional()
59
+ });
60
+ var UpdateRobotsTxtRequest = z.object({
61
+ options: z.object({
62
+ robotsTxt: z.object({
63
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
64
+ default: z.boolean().describe(
65
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
66
+ ).optional(),
67
+ subdomain: z.string().describe(
68
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
69
+ ).max(63).optional()
70
+ }).describe("`robots.txt` file to update.").optional()
71
+ }).optional()
72
+ });
73
+ var UpdateRobotsTxtResponse = z.object({
74
+ robotsTxt: z.object({
75
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
76
+ default: z.boolean().describe(
77
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
78
+ ).optional(),
79
+ subdomain: z.string().describe(
80
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
81
+ ).max(63).optional()
82
+ }).describe("Updated `robots.txt` file.").optional()
83
+ });
84
+ // Annotate the CommonJS export names for ESM import in node:
85
+ 0 && (module.exports = {
86
+ GetRobotsTxtRequest,
87
+ GetRobotsTxtResponse,
88
+ UpdateRobotsTxtRequest,
89
+ UpdateRobotsTxtResponse
90
+ });
91
+ //# sourceMappingURL=schemas.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../../schemas.ts","../../../src/promote-seo-v2-robots-txt-robots-txt.schemas.ts"],"sourcesContent":["export * from './src/promote-seo-v2-robots-txt-robots-txt.schemas.js';\n","import * as z from 'zod';\n\nexport const GetRobotsTxtRequest = z.object({\n options: z\n .object({\n subdomain: z\n .string()\n .describe(\n 'Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`.'\n )\n .optional(),\n })\n .optional(),\n});\nexport const GetRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Retrieved `robots.txt` file entity.')\n .optional(),\n});\nexport const UpdateRobotsTxtRequest = z.object({\n options: z\n .object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('`robots.txt` file to update.')\n .optional(),\n })\n .optional(),\n});\nexport const UpdateRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Updated `robots.txt` file.')\n .optional(),\n});\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,QAAmB;AAEZ,IAAM,sBAAwB,SAAO;AAAA,EAC1C,SACG,SAAO;AAAA,IACN,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,uBAAyB,SAAO;AAAA,EAC3C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,qCAAqC,EAC9C,SAAS;AACd,CAAC;AACM,IAAM,yBAA2B,SAAO;AAAA,EAC7C,SACG,SAAO;AAAA,IACN,WACG,SAAO;AAAA,MACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,MACZ,SACG,UAAQ,EACR;AAAA,QACC;AAAA,MACF,EACC,SAAS;AAAA,MACZ,WACG,SAAO,EACP;AAAA,QACC;AAAA,MACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,IACd,CAAC,EACA,SAAS,8BAA8B,EACvC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,0BAA4B,SAAO;AAAA,EAC9C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,4BAA4B,EACrC,SAAS;AACd,CAAC;","names":[]}
@@ -0,0 +1,32 @@
1
+ import * as z from 'zod';
2
+
3
+ declare const GetRobotsTxtRequest: z.ZodObject<{
4
+ options: z.ZodOptional<z.ZodObject<{
5
+ subdomain: z.ZodOptional<z.ZodString>;
6
+ }, z.core.$strip>>;
7
+ }, z.core.$strip>;
8
+ declare const GetRobotsTxtResponse: z.ZodObject<{
9
+ robotsTxt: z.ZodOptional<z.ZodObject<{
10
+ content: z.ZodOptional<z.ZodString>;
11
+ default: z.ZodOptional<z.ZodBoolean>;
12
+ subdomain: z.ZodOptional<z.ZodString>;
13
+ }, z.core.$strip>>;
14
+ }, z.core.$strip>;
15
+ declare const UpdateRobotsTxtRequest: z.ZodObject<{
16
+ options: z.ZodOptional<z.ZodObject<{
17
+ robotsTxt: z.ZodOptional<z.ZodObject<{
18
+ content: z.ZodOptional<z.ZodString>;
19
+ default: z.ZodOptional<z.ZodBoolean>;
20
+ subdomain: z.ZodOptional<z.ZodString>;
21
+ }, z.core.$strip>>;
22
+ }, z.core.$strip>>;
23
+ }, z.core.$strip>;
24
+ declare const UpdateRobotsTxtResponse: z.ZodObject<{
25
+ robotsTxt: z.ZodOptional<z.ZodObject<{
26
+ content: z.ZodOptional<z.ZodString>;
27
+ default: z.ZodOptional<z.ZodBoolean>;
28
+ subdomain: z.ZodOptional<z.ZodString>;
29
+ }, z.core.$strip>>;
30
+ }, z.core.$strip>;
31
+
32
+ export { GetRobotsTxtRequest, GetRobotsTxtResponse, UpdateRobotsTxtRequest, UpdateRobotsTxtResponse };
@@ -0,0 +1,51 @@
1
+ // src/promote-seo-v2-robots-txt-robots-txt.schemas.ts
2
+ import * as z from "zod";
3
+ var GetRobotsTxtRequest = z.object({
4
+ options: z.object({
5
+ subdomain: z.string().describe(
6
+ "Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`."
7
+ ).optional()
8
+ }).optional()
9
+ });
10
+ var GetRobotsTxtResponse = z.object({
11
+ robotsTxt: z.object({
12
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
13
+ default: z.boolean().describe(
14
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
15
+ ).optional(),
16
+ subdomain: z.string().describe(
17
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
18
+ ).max(63).optional()
19
+ }).describe("Retrieved `robots.txt` file entity.").optional()
20
+ });
21
+ var UpdateRobotsTxtRequest = z.object({
22
+ options: z.object({
23
+ robotsTxt: z.object({
24
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
25
+ default: z.boolean().describe(
26
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
27
+ ).optional(),
28
+ subdomain: z.string().describe(
29
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
30
+ ).max(63).optional()
31
+ }).describe("`robots.txt` file to update.").optional()
32
+ }).optional()
33
+ });
34
+ var UpdateRobotsTxtResponse = z.object({
35
+ robotsTxt: z.object({
36
+ content: z.string().describe("Full text content of the `robots.txt` file.").max(72e4).optional(),
37
+ default: z.boolean().describe(
38
+ "Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file)."
39
+ ).optional(),
40
+ subdomain: z.string().describe(
41
+ "If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'"
42
+ ).max(63).optional()
43
+ }).describe("Updated `robots.txt` file.").optional()
44
+ });
45
+ export {
46
+ GetRobotsTxtRequest,
47
+ GetRobotsTxtResponse,
48
+ UpdateRobotsTxtRequest,
49
+ UpdateRobotsTxtResponse
50
+ };
51
+ //# sourceMappingURL=schemas.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../../src/promote-seo-v2-robots-txt-robots-txt.schemas.ts"],"sourcesContent":["import * as z from 'zod';\n\nexport const GetRobotsTxtRequest = z.object({\n options: z\n .object({\n subdomain: z\n .string()\n .describe(\n 'Subdomain of the requested `robots.txt file`. For example, `www`, `es`, `fr`. Default: `www`.'\n )\n .optional(),\n })\n .optional(),\n});\nexport const GetRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Retrieved `robots.txt` file entity.')\n .optional(),\n});\nexport const UpdateRobotsTxtRequest = z.object({\n options: z\n .object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('`robots.txt` file to update.')\n .optional(),\n })\n .optional(),\n});\nexport const UpdateRobotsTxtResponse = z.object({\n robotsTxt: z\n .object({\n content: z\n .string()\n .describe('Full text content of the `robots.txt` file.')\n .max(720000)\n .optional(),\n default: z\n .boolean()\n .describe(\n \"Whether this `robots.txt` file uses Wix's [default content](https://support.wix.com/en/article/editing-your-sites-robotstxt-file).\"\n )\n .optional(),\n subdomain: z\n .string()\n .describe(\n \"If applicable, target subdomain for the `robots.txt` file. For example, a site can have multiple subdomains like `www.example.com`, `es.example.com`, `fr.example.com`. Each subdomain can have its own `robots.txt` file with different rules. Default: 'www'\"\n )\n .max(63)\n .optional(),\n })\n .describe('Updated `robots.txt` file.')\n .optional(),\n});\n"],"mappings":";AAAA,YAAY,OAAO;AAEZ,IAAM,sBAAwB,SAAO;AAAA,EAC1C,SACG,SAAO;AAAA,IACN,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,uBAAyB,SAAO;AAAA,EAC3C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,qCAAqC,EAC9C,SAAS;AACd,CAAC;AACM,IAAM,yBAA2B,SAAO;AAAA,EAC7C,SACG,SAAO;AAAA,IACN,WACG,SAAO;AAAA,MACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,MACZ,SACG,UAAQ,EACR;AAAA,QACC;AAAA,MACF,EACC,SAAS;AAAA,MACZ,WACG,SAAO,EACP;AAAA,QACC;AAAA,MACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,IACd,CAAC,EACA,SAAS,8BAA8B,EACvC,SAAS;AAAA,EACd,CAAC,EACA,SAAS;AACd,CAAC;AACM,IAAM,0BAA4B,SAAO;AAAA,EAC9C,WACG,SAAO;AAAA,IACN,SACG,SAAO,EACP,SAAS,6CAA6C,EACtD,IAAI,IAAM,EACV,SAAS;AAAA,IACZ,SACG,UAAQ,EACR;AAAA,MACC;AAAA,IACF,EACC,SAAS;AAAA,IACZ,WACG,SAAO,EACP;AAAA,MACC;AAAA,IACF,EACC,IAAI,EAAE,EACN,SAAS;AAAA,EACd,CAAC,EACA,SAAS,4BAA4B,EACrC,SAAS;AACd,CAAC;","names":[]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@wix/auto_sdk_seo_robots-txt",
3
- "version": "1.0.17",
3
+ "version": "1.0.19",
4
4
  "license": "MIT",
5
5
  "publishConfig": {
6
6
  "registry": "https://registry.npmjs.org/",
@@ -21,16 +21,23 @@
21
21
  "import": "./build/es/meta.mjs",
22
22
  "require": "./build/cjs/meta.js",
23
23
  "types": "./build/es/meta.d.mts"
24
+ },
25
+ "./schemas": {
26
+ "import": "./build/es/schemas.mjs",
27
+ "require": "./build/cjs/schemas.js",
28
+ "types": "./build/es/schemas.d.mts"
24
29
  }
25
30
  },
26
31
  "files": [
27
32
  "build",
28
33
  "meta",
29
- "service-plugins"
34
+ "service-plugins",
35
+ "schemas"
30
36
  ],
31
37
  "dependencies": {
32
- "@wix/sdk-runtime": "^1.0.8",
33
- "@wix/sdk-types": "^1.17.4"
38
+ "@wix/sdk-runtime": "^1.0.12",
39
+ "@wix/sdk-types": "^1.17.7",
40
+ "zod": "^4.3.6"
34
41
  },
35
42
  "devDependencies": {
36
43
  "tsup": "^8.4.0",
@@ -50,5 +57,5 @@
50
57
  "fqdn": "wix.promote.seo.v2.robots_txt"
51
58
  }
52
59
  },
53
- "falconPackageHash": "6e0e33b5e2aa34c6cad93ad49be97bd63271671026b39dcddf5eb1c6"
60
+ "falconPackageHash": "ba48b858075c51933eb09318c3c139a241d3e6a3932a95ff8e1774ce"
54
61
  }
@@ -0,0 +1,3 @@
1
+ {
2
+ "main": "../build/cjs/schemas.js"
3
+ }