@proofkit/fmdapi 5.0.3-beta.0 → 5.1.0-beta.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/intent.js +20 -0
- package/dist/esm/adapters/core.d.ts +4 -4
- package/dist/esm/adapters/fetch-base-types.d.ts +4 -4
- package/dist/esm/adapters/fetch-base.d.ts +2 -2
- package/dist/esm/adapters/fetch-base.js +36 -49
- package/dist/esm/adapters/fetch-base.js.map +1 -1
- package/dist/esm/adapters/fetch.d.ts +5 -5
- package/dist/esm/adapters/fetch.js +11 -10
- package/dist/esm/adapters/fetch.js.map +1 -1
- package/dist/esm/adapters/fm-http.d.ts +32 -0
- package/dist/esm/adapters/fm-http.js +170 -0
- package/dist/esm/adapters/fm-http.js.map +1 -0
- package/dist/esm/adapters/otto.d.ts +2 -2
- package/dist/esm/adapters/otto.js +3 -5
- package/dist/esm/adapters/otto.js.map +1 -1
- package/dist/esm/client-types.d.ts +41 -41
- package/dist/esm/client-types.js +1 -6
- package/dist/esm/client-types.js.map +1 -1
- package/dist/esm/client.d.ts +28 -44
- package/dist/esm/client.js +75 -80
- package/dist/esm/client.js.map +1 -1
- package/dist/esm/index.d.ts +5 -6
- package/dist/esm/index.js +7 -5
- package/dist/esm/index.js.map +1 -1
- package/dist/esm/tokenStore/index.d.ts +1 -1
- package/dist/esm/tokenStore/memory.js.map +1 -1
- package/dist/esm/tokenStore/types.d.ts +2 -2
- package/dist/esm/tokenStore/upstash.d.ts +1 -1
- package/dist/esm/utils.d.ts +7 -7
- package/dist/esm/utils.js +6 -4
- package/dist/esm/utils.js.map +1 -1
- package/package.json +37 -26
- package/skills/fmdapi-client/SKILL.md +490 -0
- package/src/adapters/core.ts +6 -9
- package/src/adapters/fetch-base-types.ts +5 -3
- package/src/adapters/fetch-base.ts +53 -78
- package/src/adapters/fetch.ts +19 -24
- package/src/adapters/fm-http.ts +224 -0
- package/src/adapters/otto.ts +8 -8
- package/src/client-types.ts +59 -83
- package/src/client.ts +131 -167
- package/src/index.ts +5 -9
- package/src/tokenStore/file.ts +2 -4
- package/src/tokenStore/index.ts +1 -1
- package/src/tokenStore/types.ts +2 -2
- package/src/tokenStore/upstash.ts +2 -5
- package/src/utils.ts +16 -23
package/dist/esm/client.js
CHANGED
|
@@ -1,12 +1,10 @@
|
|
|
1
1
|
import { FileMakerError } from "./client-types.js";
|
|
2
2
|
function asNumber(input) {
|
|
3
|
-
return typeof input === "string" ? parseInt(input) : input;
|
|
3
|
+
return typeof input === "string" ? Number.parseInt(input, 10) : input;
|
|
4
4
|
}
|
|
5
5
|
function DataApi(options) {
|
|
6
6
|
if ("zodValidators" in options) {
|
|
7
|
-
throw new Error(
|
|
8
|
-
"zodValidators is no longer supported. Use schema instead, or re-run the typegen command"
|
|
9
|
-
);
|
|
7
|
+
throw new Error("zodValidators is no longer supported. Use schema instead, or re-run the typegen command");
|
|
10
8
|
}
|
|
11
9
|
const schema = options.schema;
|
|
12
10
|
const layout = options.layout;
|
|
@@ -24,33 +22,33 @@ function DataApi(options) {
|
|
|
24
22
|
} = options.adapter;
|
|
25
23
|
async function _list(args) {
|
|
26
24
|
const { fetch, timeout, ...params } = args ?? {};
|
|
27
|
-
if ("limit" in params && params.limit !== void 0)
|
|
28
|
-
|
|
25
|
+
if ("limit" in params && params.limit !== void 0) {
|
|
26
|
+
Object.assign(params, { _limit: params.limit }).limit = void 0;
|
|
27
|
+
}
|
|
29
28
|
if ("offset" in params && params.offset !== void 0) {
|
|
30
|
-
if (params.offset <= 1)
|
|
31
|
-
|
|
29
|
+
if (params.offset <= 1) {
|
|
30
|
+
params.offset = void 0;
|
|
31
|
+
} else {
|
|
32
|
+
Object.assign(params, { _offset: params.offset }).offset = void 0;
|
|
33
|
+
}
|
|
32
34
|
}
|
|
33
|
-
if ("sort" in params && params.sort !== void 0)
|
|
34
|
-
|
|
35
|
+
if ("sort" in params && params.sort !== void 0) {
|
|
36
|
+
Object.assign(params, {
|
|
35
37
|
_sort: Array.isArray(params.sort) ? params.sort : [params.sort]
|
|
36
|
-
})
|
|
38
|
+
}).sort = void 0;
|
|
39
|
+
}
|
|
37
40
|
const result = await list({
|
|
38
41
|
layout,
|
|
39
42
|
data: params,
|
|
40
43
|
fetch,
|
|
41
44
|
timeout
|
|
42
45
|
});
|
|
43
|
-
if (result.dataInfo.foundCount > result.dataInfo.returnedCount) {
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
);
|
|
48
|
-
}
|
|
46
|
+
if (result.dataInfo.foundCount > result.dataInfo.returnedCount && (args == null ? void 0 : args.limit) === void 0 && (args == null ? void 0 : args.offset) === void 0) {
|
|
47
|
+
console.warn(
|
|
48
|
+
`🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your "${layout}" layout. Use the "listAll" method to automatically paginate through all records, or specify a "limit" and "offset" to handle pagination yourself.`
|
|
49
|
+
);
|
|
49
50
|
}
|
|
50
|
-
return await runSchemaValidationAndTransform(
|
|
51
|
-
schema,
|
|
52
|
-
result
|
|
53
|
-
);
|
|
51
|
+
return await runSchemaValidationAndTransform(schema, result);
|
|
54
52
|
}
|
|
55
53
|
async function listAll(args) {
|
|
56
54
|
let runningData = [];
|
|
@@ -60,11 +58,12 @@ function DataApi(options) {
|
|
|
60
58
|
const data = await _list({
|
|
61
59
|
...args,
|
|
62
60
|
offset
|
|
63
|
-
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
64
61
|
});
|
|
65
62
|
runningData = [...runningData, ...data.data];
|
|
66
|
-
if (runningData.length >= data.dataInfo.foundCount)
|
|
67
|
-
|
|
63
|
+
if (runningData.length >= data.dataInfo.foundCount) {
|
|
64
|
+
break;
|
|
65
|
+
}
|
|
66
|
+
offset += limit;
|
|
68
67
|
}
|
|
69
68
|
return runningData;
|
|
70
69
|
}
|
|
@@ -86,10 +85,7 @@ function DataApi(options) {
|
|
|
86
85
|
fetch,
|
|
87
86
|
timeout
|
|
88
87
|
});
|
|
89
|
-
return await runSchemaValidationAndTransform(
|
|
90
|
-
schema,
|
|
91
|
-
result
|
|
92
|
-
);
|
|
88
|
+
return await runSchemaValidationAndTransform(schema, result);
|
|
93
89
|
}
|
|
94
90
|
async function _update(args) {
|
|
95
91
|
args.recordId = asNumber(args.recordId);
|
|
@@ -101,7 +97,7 @@ function DataApi(options) {
|
|
|
101
97
|
timeout
|
|
102
98
|
});
|
|
103
99
|
}
|
|
104
|
-
|
|
100
|
+
function deleteRecord(args) {
|
|
105
101
|
args.recordId = asNumber(args.recordId);
|
|
106
102
|
const { recordId, fetch, timeout, ...params } = args;
|
|
107
103
|
return _adapterDelete({
|
|
@@ -112,19 +108,23 @@ function DataApi(options) {
|
|
|
112
108
|
});
|
|
113
109
|
}
|
|
114
110
|
async function _find(args) {
|
|
115
|
-
const {
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
fetch,
|
|
120
|
-
...params
|
|
121
|
-
} = args;
|
|
122
|
-
const query = !Array.isArray(queryInput) ? [queryInput] : queryInput;
|
|
123
|
-
if ("offset" in params && params.offset !== void 0) {
|
|
124
|
-
if (params.offset <= 1) delete params.offset;
|
|
111
|
+
const { query: queryInput, ignoreEmptyResult = false, timeout, fetch, ...params } = args;
|
|
112
|
+
const query = Array.isArray(queryInput) ? queryInput : [queryInput];
|
|
113
|
+
if ("offset" in params && params.offset !== void 0 && params.offset <= 1) {
|
|
114
|
+
params.offset = void 0;
|
|
125
115
|
}
|
|
126
116
|
if ("dateformats" in params && params.dateformats !== void 0) {
|
|
127
|
-
|
|
117
|
+
let dateFormatValue;
|
|
118
|
+
if (params.dateformats === "US") {
|
|
119
|
+
dateFormatValue = 0;
|
|
120
|
+
} else if (params.dateformats === "file_locale") {
|
|
121
|
+
dateFormatValue = 1;
|
|
122
|
+
} else if (params.dateformats === "ISO8601") {
|
|
123
|
+
dateFormatValue = 2;
|
|
124
|
+
} else {
|
|
125
|
+
dateFormatValue = 0;
|
|
126
|
+
}
|
|
127
|
+
params.dateformats = dateFormatValue.toString();
|
|
128
128
|
}
|
|
129
129
|
const result = await find({
|
|
130
130
|
data: { ...params, query },
|
|
@@ -132,48 +132,43 @@ function DataApi(options) {
|
|
|
132
132
|
fetch,
|
|
133
133
|
timeout
|
|
134
134
|
}).catch((e) => {
|
|
135
|
-
if (ignoreEmptyResult && e instanceof FileMakerError && e.code === "401")
|
|
135
|
+
if (ignoreEmptyResult && e instanceof FileMakerError && e.code === "401") {
|
|
136
136
|
return { data: [], dataInfo: { foundCount: 0, returnedCount: 0 } };
|
|
137
|
+
}
|
|
137
138
|
throw e;
|
|
138
139
|
});
|
|
139
|
-
if (result.dataInfo.foundCount > result.dataInfo.returnedCount) {
|
|
140
|
-
|
|
141
|
-
|
|
142
|
-
|
|
143
|
-
);
|
|
144
|
-
}
|
|
140
|
+
if (result.dataInfo.foundCount > result.dataInfo.returnedCount && (args == null ? void 0 : args.limit) === void 0 && (args == null ? void 0 : args.offset) === void 0) {
|
|
141
|
+
console.warn(
|
|
142
|
+
`🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your "${layout}" layout. Use the "findAll" method to automatically paginate through all records, or specify a "limit" and "offset" to handle pagination yourself.`
|
|
143
|
+
);
|
|
145
144
|
}
|
|
146
145
|
return await runSchemaValidationAndTransform(schema, result);
|
|
147
146
|
}
|
|
148
147
|
async function findOne(args) {
|
|
149
148
|
const result = await _find(args);
|
|
150
|
-
if (result.data.length !== 1)
|
|
151
|
-
throw new Error(
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
|
|
156
|
-
|
|
157
|
-
);
|
|
158
|
-
if (!transformedResult.data[0]) throw new Error("No data found");
|
|
149
|
+
if (result.data.length !== 1) {
|
|
150
|
+
throw new Error(`${result.data.length} records found; expecting exactly 1`);
|
|
151
|
+
}
|
|
152
|
+
const transformedResult = await runSchemaValidationAndTransform(schema, result);
|
|
153
|
+
if (!transformedResult.data[0]) {
|
|
154
|
+
throw new Error("No data found");
|
|
155
|
+
}
|
|
159
156
|
return { ...transformedResult, data: transformedResult.data[0] };
|
|
160
157
|
}
|
|
161
158
|
async function findFirst(args) {
|
|
162
159
|
const result = await _find(args);
|
|
163
|
-
const transformedResult = await runSchemaValidationAndTransform(
|
|
164
|
-
|
|
165
|
-
|
|
166
|
-
|
|
167
|
-
if (!transformedResult.data[0]) throw new Error("No data found");
|
|
160
|
+
const transformedResult = await runSchemaValidationAndTransform(schema, result);
|
|
161
|
+
if (!transformedResult.data[0]) {
|
|
162
|
+
throw new Error("No data found");
|
|
163
|
+
}
|
|
168
164
|
return { ...transformedResult, data: transformedResult.data[0] };
|
|
169
165
|
}
|
|
170
166
|
async function maybeFindFirst(args) {
|
|
171
167
|
const result = await _find({ ...args, ignoreEmptyResult: true });
|
|
172
|
-
const transformedResult = await runSchemaValidationAndTransform(
|
|
173
|
-
|
|
174
|
-
|
|
175
|
-
|
|
176
|
-
if (!transformedResult.data[0]) return null;
|
|
168
|
+
const transformedResult = await runSchemaValidationAndTransform(schema, result);
|
|
169
|
+
if (!transformedResult.data[0]) {
|
|
170
|
+
return null;
|
|
171
|
+
}
|
|
177
172
|
return { ...transformedResult, data: transformedResult.data[0] };
|
|
178
173
|
}
|
|
179
174
|
async function findAll(args) {
|
|
@@ -187,9 +182,10 @@ function DataApi(options) {
|
|
|
187
182
|
ignoreEmptyResult: true
|
|
188
183
|
});
|
|
189
184
|
runningData = [...runningData, ...data.data];
|
|
190
|
-
if (runningData.length === 0 || runningData.length >= data.dataInfo.foundCount)
|
|
185
|
+
if (runningData.length === 0 || runningData.length >= data.dataInfo.foundCount) {
|
|
191
186
|
break;
|
|
192
|
-
|
|
187
|
+
}
|
|
188
|
+
offset += limit;
|
|
193
189
|
}
|
|
194
190
|
return runningData;
|
|
195
191
|
}
|
|
@@ -221,29 +217,28 @@ function DataApi(options) {
|
|
|
221
217
|
var _a;
|
|
222
218
|
const fieldDataIssues = [];
|
|
223
219
|
const portalDataIssues = [];
|
|
224
|
-
if (!schema2)
|
|
220
|
+
if (!schema2) {
|
|
221
|
+
return result;
|
|
222
|
+
}
|
|
225
223
|
const transformedData = [];
|
|
226
224
|
for (const record of result.data) {
|
|
227
|
-
let fieldResult = schema2.fieldData["~standard"].validate(
|
|
228
|
-
|
|
229
|
-
|
|
230
|
-
|
|
225
|
+
let fieldResult = schema2.fieldData["~standard"].validate(record.fieldData);
|
|
226
|
+
if (fieldResult instanceof Promise) {
|
|
227
|
+
fieldResult = await fieldResult;
|
|
228
|
+
}
|
|
231
229
|
if ("value" in fieldResult) {
|
|
232
230
|
record.fieldData = fieldResult.value;
|
|
233
231
|
} else {
|
|
234
232
|
fieldDataIssues.push(...fieldResult.issues);
|
|
235
233
|
}
|
|
236
234
|
if (schema2.portalData) {
|
|
237
|
-
for (const [portalName, portalRecords] of Object.entries(
|
|
238
|
-
record.portalData
|
|
239
|
-
)) {
|
|
235
|
+
for (const [portalName, portalRecords] of Object.entries(record.portalData)) {
|
|
240
236
|
const validatedPortalRecords = [];
|
|
241
237
|
for (const portalRecord of portalRecords) {
|
|
242
|
-
let portalResult = (_a = schema2.portalData[portalName]) == null ? void 0 : _a["~standard"].validate(
|
|
243
|
-
|
|
244
|
-
);
|
|
245
|
-
if (portalResult instanceof Promise)
|
|
238
|
+
let portalResult = (_a = schema2.portalData[portalName]) == null ? void 0 : _a["~standard"].validate(portalRecord);
|
|
239
|
+
if (portalResult instanceof Promise) {
|
|
246
240
|
portalResult = await portalResult;
|
|
241
|
+
}
|
|
247
242
|
if (portalResult && "value" in portalResult) {
|
|
248
243
|
validatedPortalRecords.push({
|
|
249
244
|
...portalResult.value,
|
package/dist/esm/client.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"client.js","sources":["../../src/client.ts"],"sourcesContent":["import type { Adapter, ExecuteScriptOptions } from \"./adapters/core.js\";\nimport type {\n CreateParams,\n CreateResponse,\n DeleteParams,\n DeleteResponse,\n FMRecord,\n FieldData,\n GenericPortalData,\n GetParams,\n GetResponse,\n GetResponseOne,\n ListParams,\n PortalsWithIds,\n Query,\n UpdateParams,\n UpdateResponse,\n} from \"./client-types.js\";\nimport { FileMakerError } from \"./index.js\";\nimport type { StandardSchemaV1 } from \"@standard-schema/spec\";\n\nfunction asNumber(input: string | number): number {\n return typeof input === \"string\" ? parseInt(input) : input;\n}\n\nexport type ClientObjectProps = {\n /**\n * The layout to use by default for all requests. Can be overrridden on each request.\n */\n layout: string;\n schema?: {\n /**\n * The schema for the field data.\n */\n fieldData: StandardSchemaV1<FieldData>;\n /**\n * The schema for the portal data.\n */\n portalData?: Record<string, StandardSchemaV1<FieldData>>;\n };\n};\n\ntype FetchOptions = {\n fetch?: RequestInit;\n};\n\nfunction DataApi<\n Fd extends FieldData = FieldData,\n Pd extends GenericPortalData = GenericPortalData,\n Opts extends ClientObjectProps = ClientObjectProps,\n Adp extends Adapter = Adapter,\n>(options: Opts & { adapter: Adp }) {\n type InferredFieldData = Opts[\"schema\"] extends object\n ? StandardSchemaV1.InferOutput<Opts[\"schema\"][\"fieldData\"]>\n : Fd;\n type InferredPortalData = Opts[\"schema\"] extends object\n ? Opts[\"schema\"][\"portalData\"] extends object\n ? {\n [K in keyof Opts[\"schema\"][\"portalData\"]]: StandardSchemaV1.InferOutput<\n Opts[\"schema\"][\"portalData\"][K]\n >;\n }\n : Pd\n : Pd;\n\n if (\"zodValidators\" in options) {\n throw new Error(\n \"zodValidators is no longer supported. Use schema instead, or re-run the typegen command\",\n );\n }\n\n const schema = options.schema;\n const layout = options.layout;\n const {\n create,\n delete: _adapterDelete,\n find,\n get,\n list,\n update,\n layoutMetadata,\n containerUpload,\n executeScript,\n ...otherMethods\n } = options.adapter;\n\n type CreateArgs<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = CreateParams<U> & {\n fieldData: Partial<T>;\n };\n type GetArgs<U extends InferredPortalData = InferredPortalData> =\n GetParams<U> & {\n recordId: number | string;\n };\n type UpdateArgs<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = UpdateParams<U> & {\n fieldData: Partial<T>;\n recordId: number | string;\n };\n type ContainerUploadArgs<T extends InferredFieldData = InferredFieldData> = {\n containerFieldName: keyof T;\n containerFieldRepetition?: string | number;\n file: Blob;\n recordId: number | string;\n modId?: number;\n timeout?: number;\n };\n type DeleteArgs = DeleteParams & {\n recordId: number | string;\n };\n type IgnoreEmptyResult = {\n /**\n * If true, a find that returns no results will retun an empty array instead of throwing an error.\n * @default false\n */\n ignoreEmptyResult?: boolean;\n };\n type FindArgs<\n T extends FieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = ListParams<T, U> & {\n query: Query<T> | Array<Query<T>>;\n timeout?: number;\n };\n\n type ExecuteScriptArgs = Omit<ExecuteScriptOptions, \"layout\">;\n\n /**\n * List all records from a given layout, no find criteria applied.\n */\n async function _list(): Promise<\n GetResponse<InferredFieldData, InferredPortalData>\n >;\n async function _list(\n args: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>>;\n async function _list(\n args?: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const { fetch, timeout, ...params } = args ?? {};\n\n // rename and refactor limit, offset, and sort keys for this request\n if (\"limit\" in params && params.limit !== undefined)\n delete Object.assign(params, { _limit: params.limit })[\"limit\"];\n if (\"offset\" in params && params.offset !== undefined) {\n if (params.offset <= 1) delete params.offset;\n else delete Object.assign(params, { _offset: params.offset })[\"offset\"];\n }\n if (\"sort\" in params && params.sort !== undefined)\n delete Object.assign(params, {\n _sort: Array.isArray(params.sort) ? params.sort : [params.sort],\n })[\"sort\"];\n\n const result = await list({\n layout,\n data: params,\n fetch,\n timeout,\n });\n\n if (result.dataInfo.foundCount > result.dataInfo.returnedCount) {\n // more records found than returned\n if (args?.limit === undefined && args?.offset === undefined) {\n // and the user didn't specify a limit or offset, so we should warn them\n console.warn(\n `🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your \"${layout}\" layout. Use the \"listAll\" method to automatically paginate through all records, or specify a \"limit\" and \"offset\" to handle pagination yourself.`,\n );\n }\n }\n\n return await runSchemaValidationAndTransform(\n schema,\n result as GetResponse<InferredFieldData, InferredPortalData>,\n );\n }\n\n /**\n * Paginate through all records from a given layout, no find criteria applied.\n * ⚠️ WARNING: Use this method with caution, as it can be slow with large datasets\n */\n async function listAll<\n T extends FieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n >(): Promise<FMRecord<T, U>[]>;\n async function listAll<\n T extends FieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n >(args: ListParams<T, U> & FetchOptions): Promise<FMRecord<T, U>[]>;\n async function listAll<\n T extends FieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n >(args?: ListParams<T, U> & FetchOptions): Promise<FMRecord<T, U>[]> {\n let runningData: GetResponse<T, U>[\"data\"] = [];\n const limit = args?.limit ?? 100;\n let offset = args?.offset ?? 1;\n\n // eslint-disable-next-line no-constant-condition\n while (true) {\n const data = (await _list({\n ...args,\n offset,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n } as any)) as unknown as GetResponse<T, U>;\n runningData = [...runningData, ...data.data];\n if (runningData.length >= data.dataInfo.foundCount) break;\n offset = offset + limit;\n }\n return runningData;\n }\n\n /**\n * Create a new record in a given layout\n */\n async function _create<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n >(args: CreateArgs<T, U> & FetchOptions): Promise<CreateResponse> {\n const { fetch, timeout, ...params } = args ?? {};\n return await create({\n layout,\n data: params,\n fetch,\n timeout,\n });\n }\n\n /**\n * Get a single record by Internal RecordId\n */\n async function _get(\n args: GetArgs<InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n\n const result = await get({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n return await runSchemaValidationAndTransform(\n schema,\n result as GetResponse<InferredFieldData, InferredPortalData>,\n );\n }\n\n /**\n * Update a single record by internal RecordId\n */\n async function _update(\n args: UpdateArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<UpdateResponse> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n return await update({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n }\n\n /**\n * Delete a single record by internal RecordId\n */\n async function deleteRecord(\n args: DeleteArgs & FetchOptions,\n ): Promise<DeleteResponse> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n\n return _adapterDelete({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n }\n\n /**\n * Find records in a given layout\n */\n async function _find(\n args: FindArgs<InferredFieldData, InferredPortalData> &\n IgnoreEmptyResult &\n FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const {\n query: queryInput,\n ignoreEmptyResult = false,\n timeout,\n fetch,\n ...params\n } = args;\n const query = !Array.isArray(queryInput) ? [queryInput] : queryInput;\n\n // rename and refactor limit, offset, and sort keys for this request\n if (\"offset\" in params && params.offset !== undefined) {\n if (params.offset <= 1) delete params.offset;\n }\n if (\"dateformats\" in params && params.dateformats !== undefined) {\n // reassign dateformats to match FileMaker's expected values\n // @ts-expect-error FM wants a string, so this is fine\n params.dateformats = (\n params.dateformats === \"US\"\n ? 0\n : params.dateformats === \"file_locale\"\n ? 1\n : params.dateformats === \"ISO8601\"\n ? 2\n : 0\n ).toString();\n }\n const result = (await find({\n data: { ...params, query },\n layout,\n fetch,\n timeout,\n }).catch((e: unknown) => {\n if (ignoreEmptyResult && e instanceof FileMakerError && e.code === \"401\")\n return { data: [], dataInfo: { foundCount: 0, returnedCount: 0 } };\n throw e;\n })) as GetResponse<InferredFieldData, InferredPortalData>;\n\n if (result.dataInfo.foundCount > result.dataInfo.returnedCount) {\n // more records found than returned\n if (args?.limit === undefined && args?.offset === undefined) {\n console.warn(\n `🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your \"${layout}\" layout. Use the \"findAll\" method to automatically paginate through all records, or specify a \"limit\" and \"offset\" to handle pagination yourself.`,\n );\n }\n }\n\n return await runSchemaValidationAndTransform(schema, result);\n }\n\n /**\n * Helper method for `find`. Will only return the first result or throw error if there is more than 1 result.\n */\n async function findOne(\n args: FindArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData>> {\n const result = await _find(args);\n if (result.data.length !== 1)\n throw new Error(\n `${result.data.length} records found; expecting exactly 1`,\n );\n const transformedResult = await runSchemaValidationAndTransform(\n schema,\n result,\n );\n if (!transformedResult.data[0]) throw new Error(\"No data found\");\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find`. Will only return the first result instead of an array.\n */\n async function findFirst(\n args: FindArgs<InferredFieldData, InferredPortalData> &\n IgnoreEmptyResult &\n FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData>> {\n const result = await _find(args);\n const transformedResult = await runSchemaValidationAndTransform(\n schema,\n result,\n );\n\n if (!transformedResult.data[0]) throw new Error(\"No data found\");\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find`. Will return the first result or null if no results are found.\n */\n async function maybeFindFirst(\n args: FindArgs<InferredFieldData, InferredPortalData> &\n IgnoreEmptyResult &\n FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData> | null> {\n const result = await _find({ ...args, ignoreEmptyResult: true });\n const transformedResult = await runSchemaValidationAndTransform(\n schema,\n result,\n );\n if (!transformedResult.data[0]) return null;\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find` to page through all found results.\n * ⚠️ WARNING: Use with caution as this can be a slow operation with large datasets\n */\n async function findAll(\n args: FindArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<FMRecord<InferredFieldData, InferredPortalData>[]> {\n let runningData: GetResponse<\n InferredFieldData,\n InferredPortalData\n >[\"data\"] = [];\n const limit = args.limit ?? 100;\n let offset = args.offset ?? 1;\n // eslint-disable-next-line no-constant-condition\n while (true) {\n const data = await _find({\n ...args,\n offset,\n ignoreEmptyResult: true,\n });\n runningData = [...runningData, ...data.data];\n if (\n runningData.length === 0 ||\n runningData.length >= data.dataInfo.foundCount\n )\n break;\n offset = offset + limit;\n }\n return runningData;\n }\n\n async function _layoutMetadata(args?: { timeout?: number } & FetchOptions) {\n const { ...restArgs } = args ?? {};\n // Explicitly define the type for params based on FetchOptions\n const params: FetchOptions & { timeout?: number } = restArgs;\n\n return await layoutMetadata({\n layout,\n fetch: params.fetch, // Now should correctly resolve to undefined if not present\n timeout: params.timeout, // Now should correctly resolve to undefined if not present\n });\n }\n\n async function _containerUpload<\n T extends InferredFieldData = InferredFieldData,\n >(args: ContainerUploadArgs<T> & FetchOptions) {\n const { ...params } = args;\n return await containerUpload({\n layout,\n data: {\n ...params,\n containerFieldName: params.containerFieldName as string,\n repetition: params.containerFieldRepetition,\n },\n fetch: params.fetch,\n timeout: params.timeout,\n });\n }\n\n async function runSchemaValidationAndTransform(\n schema: ClientObjectProps[\"schema\"],\n result: GetResponse<InferredFieldData, InferredPortalData>,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const fieldDataIssues: StandardSchemaV1.Issue[] = [];\n const portalDataIssues: StandardSchemaV1.Issue[] = [];\n\n if (!schema) return result;\n const transformedData: FMRecord<InferredFieldData, InferredPortalData>[] =\n [];\n for (const record of result.data) {\n let fieldResult = schema.fieldData[\"~standard\"].validate(\n record.fieldData,\n );\n if (fieldResult instanceof Promise) fieldResult = await fieldResult;\n if (\"value\" in fieldResult) {\n record.fieldData = fieldResult.value as InferredFieldData;\n } else {\n fieldDataIssues.push(...fieldResult.issues);\n }\n\n if (schema.portalData) {\n for (const [portalName, portalRecords] of Object.entries(\n record.portalData,\n )) {\n const validatedPortalRecords: PortalsWithIds<GenericPortalData>[] =\n [];\n for (const portalRecord of portalRecords) {\n let portalResult =\n schema.portalData[portalName]?.[\"~standard\"].validate(\n portalRecord,\n );\n if (portalResult instanceof Promise)\n portalResult = await portalResult;\n if (portalResult && \"value\" in portalResult) {\n validatedPortalRecords.push({\n ...portalResult.value,\n recordId: portalRecord.recordId,\n modId: portalRecord.modId,\n });\n } else {\n portalDataIssues.push(...(portalResult?.issues ?? []));\n }\n }\n // @ts-expect-error We know portalName is a valid key, but can't figure out the right assertions\n record.portalData[portalName] = validatedPortalRecords;\n }\n }\n\n transformedData.push(record);\n }\n result.data = transformedData;\n\n if (fieldDataIssues.length > 0 || portalDataIssues.length > 0) {\n console.error(\n `🚨 @proofkit/fmdapi: Validation issues for layout \"${layout}\". Run the typegen command again to generate the latest field definitions from your layout.`,\n {\n fieldDataIssues,\n portalDataIssues,\n },\n );\n throw new Error(\"Schema validation issues\");\n }\n\n return result;\n }\n\n async function _executeScript(args: ExecuteScriptArgs & FetchOptions) {\n return await executeScript({\n ...args,\n layout,\n });\n }\n\n return {\n ...otherMethods,\n layout: options.layout as Opts[\"layout\"],\n list: _list,\n listAll,\n create: _create,\n get: _get,\n update: _update,\n delete: deleteRecord,\n find: _find,\n findOne,\n findFirst,\n maybeFindFirst,\n findAll,\n layoutMetadata: _layoutMetadata,\n containerUpload: _containerUpload,\n executeScript: _executeScript,\n };\n}\n\nexport default DataApi;\nexport { DataApi };\n"],"names":["schema"],"mappings":";AAqBA,SAAS,SAAS,OAAgC;AAChD,SAAO,OAAO,UAAU,WAAW,SAAS,KAAK,IAAI;AACvD;AAuBA,SAAS,QAKP,SAAkC;AAclC,MAAI,mBAAmB,SAAS;AAC9B,UAAM,IAAI;AAAA,MACR;AAAA,IACF;AAAA,EAAA;AAGF,QAAM,SAAS,QAAQ;AACvB,QAAM,SAAS,QAAQ;AACjB,QAAA;AAAA,IACJ;AAAA,IACA,QAAQ;AAAA,IACR;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,GAAG;AAAA,MACD,QAAQ;AAwDZ,iBAAe,MACb,MAC6D;AAC7D,UAAM,EAAE,OAAO,SAAS,GAAG,OAAO,IAAI,QAAQ,CAAC;AAG3C,QAAA,WAAW,UAAU,OAAO,UAAU;AACjC,aAAA,OAAO,OAAO,QAAQ,EAAE,QAAQ,OAAO,OAAO,EAAE,OAAO;AAChE,QAAI,YAAY,UAAU,OAAO,WAAW,QAAW;AACrD,UAAI,OAAO,UAAU,EAAG,QAAO,OAAO;AAAA,UACjC,QAAO,OAAO,OAAO,QAAQ,EAAE,SAAS,OAAO,QAAQ,EAAE,QAAQ;AAAA,IAAA;AAEpE,QAAA,UAAU,UAAU,OAAO,SAAS;AAC/B,aAAA,OAAO,OAAO,QAAQ;AAAA,QAC3B,OAAO,MAAM,QAAQ,OAAO,IAAI,IAAI,OAAO,OAAO,CAAC,OAAO,IAAI;AAAA,MAC/D,CAAA,EAAE,MAAM;AAEL,UAAA,SAAS,MAAM,KAAK;AAAA,MACxB;AAAA,MACA,MAAM;AAAA,MACN;AAAA,MACA;AAAA,IAAA,CACD;AAED,QAAI,OAAO,SAAS,aAAa,OAAO,SAAS,eAAe;AAE9D,WAAI,6BAAM,WAAU,WAAa,6BAAM,YAAW,QAAW;AAEnD,gBAAA;AAAA,UACN,oCAAoC,OAAO,SAAS,aAAa,WAAW,OAAO,SAAS,UAAU,uBAAuB,MAAM;AAAA,QACrI;AAAA,MAAA;AAAA,IACF;AAGF,WAAO,MAAM;AAAA,MACX;AAAA,MACA;AAAA,IACF;AAAA,EAAA;AAeF,iBAAe,QAGb,MAAmE;AACnE,QAAI,cAAyC,CAAC;AACxC,UAAA,SAAQ,6BAAM,UAAS;AACzB,QAAA,UAAS,6BAAM,WAAU;AAG7B,WAAO,MAAM;AACL,YAAA,OAAQ,MAAM,MAAM;AAAA,QACxB,GAAG;AAAA,QACH;AAAA;AAAA,MAAA,CAEM;AACR,oBAAc,CAAC,GAAG,aAAa,GAAG,KAAK,IAAI;AAC3C,UAAI,YAAY,UAAU,KAAK,SAAS,WAAY;AACpD,eAAS,SAAS;AAAA,IAAA;AAEb,WAAA;AAAA,EAAA;AAMT,iBAAe,QAGb,MAAgE;AAChE,UAAM,EAAE,OAAO,SAAS,GAAG,OAAO,IAAI,QAAQ,CAAC;AAC/C,WAAO,MAAM,OAAO;AAAA,MAClB;AAAA,MACA,MAAM;AAAA,MACN;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EAAA;AAMH,iBAAe,KACb,MAC6D;AACxD,SAAA,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,OAAW,IAAA;AAE1C,UAAA,SAAS,MAAM,IAAI;AAAA,MACvB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAS;AAAA,MAC5B;AAAA,MACA;AAAA,IAAA,CACD;AACD,WAAO,MAAM;AAAA,MACX;AAAA,MACA;AAAA,IACF;AAAA,EAAA;AAMF,iBAAe,QACb,MACyB;AACpB,SAAA,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,OAAW,IAAA;AAChD,WAAO,MAAM,OAAO;AAAA,MAClB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAS;AAAA,MAC5B;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EAAA;AAMH,iBAAe,aACb,MACyB;AACpB,SAAA,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,OAAW,IAAA;AAEhD,WAAO,eAAe;AAAA,MACpB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAS;AAAA,MAC5B;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EAAA;AAMH,iBAAe,MACb,MAG6D;AACvD,UAAA;AAAA,MACJ,OAAO;AAAA,MACP,oBAAoB;AAAA,MACpB;AAAA,MACA;AAAA,MACA,GAAG;AAAA,IAAA,IACD;AACE,UAAA,QAAQ,CAAC,MAAM,QAAQ,UAAU,IAAI,CAAC,UAAU,IAAI;AAG1D,QAAI,YAAY,UAAU,OAAO,WAAW,QAAW;AACrD,UAAI,OAAO,UAAU,EAAG,QAAO,OAAO;AAAA,IAAA;AAExC,QAAI,iBAAiB,UAAU,OAAO,gBAAgB,QAAW;AAG/D,aAAO,eACL,OAAO,gBAAgB,OACnB,IACA,OAAO,gBAAgB,gBACrB,IACA,OAAO,gBAAgB,YACrB,IACA,GACR,SAAS;AAAA,IAAA;AAEP,UAAA,SAAU,MAAM,KAAK;AAAA,MACzB,MAAM,EAAE,GAAG,QAAQ,MAAM;AAAA,MACzB;AAAA,MACA;AAAA,MACA;AAAA,IAAA,CACD,EAAE,MAAM,CAAC,MAAe;AACvB,UAAI,qBAAqB,aAAa,kBAAkB,EAAE,SAAS;AAC1D,eAAA,EAAE,MAAM,CAAA,GAAI,UAAU,EAAE,YAAY,GAAG,eAAe,IAAI;AAC7D,YAAA;AAAA,IAAA,CACP;AAED,QAAI,OAAO,SAAS,aAAa,OAAO,SAAS,eAAe;AAE9D,WAAI,6BAAM,WAAU,WAAa,6BAAM,YAAW,QAAW;AACnD,gBAAA;AAAA,UACN,oCAAoC,OAAO,SAAS,aAAa,WAAW,OAAO,SAAS,UAAU,uBAAuB,MAAM;AAAA,QACrI;AAAA,MAAA;AAAA,IACF;AAGK,WAAA,MAAM,gCAAgC,QAAQ,MAAM;AAAA,EAAA;AAM7D,iBAAe,QACb,MACgE;AAC1D,UAAA,SAAS,MAAM,MAAM,IAAI;AAC3B,QAAA,OAAO,KAAK,WAAW;AACzB,YAAM,IAAI;AAAA,QACR,GAAG,OAAO,KAAK,MAAM;AAAA,MACvB;AACF,UAAM,oBAAoB,MAAM;AAAA,MAC9B;AAAA,MACA;AAAA,IACF;AACI,QAAA,CAAC,kBAAkB,KAAK,CAAC,EAAS,OAAA,IAAI,MAAM,eAAe;AAC/D,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAE;AAAA,EAAA;AAMjE,iBAAe,UACb,MAGgE;AAC1D,UAAA,SAAS,MAAM,MAAM,IAAI;AAC/B,UAAM,oBAAoB,MAAM;AAAA,MAC9B;AAAA,MACA;AAAA,IACF;AAEI,QAAA,CAAC,kBAAkB,KAAK,CAAC,EAAS,OAAA,IAAI,MAAM,eAAe;AAC/D,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAE;AAAA,EAAA;AAMjE,iBAAe,eACb,MAGuE;AACjE,UAAA,SAAS,MAAM,MAAM,EAAE,GAAG,MAAM,mBAAmB,MAAM;AAC/D,UAAM,oBAAoB,MAAM;AAAA,MAC9B;AAAA,MACA;AAAA,IACF;AACA,QAAI,CAAC,kBAAkB,KAAK,CAAC,EAAU,QAAA;AACvC,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAE;AAAA,EAAA;AAOjE,iBAAe,QACb,MAC4D;AAC5D,QAAI,cAGQ,CAAC;AACP,UAAA,QAAQ,KAAK,SAAS;AACxB,QAAA,SAAS,KAAK,UAAU;AAE5B,WAAO,MAAM;AACL,YAAA,OAAO,MAAM,MAAM;AAAA,QACvB,GAAG;AAAA,QACH;AAAA,QACA,mBAAmB;AAAA,MAAA,CACpB;AACD,oBAAc,CAAC,GAAG,aAAa,GAAG,KAAK,IAAI;AAC3C,UACE,YAAY,WAAW,KACvB,YAAY,UAAU,KAAK,SAAS;AAEpC;AACF,eAAS,SAAS;AAAA,IAAA;AAEb,WAAA;AAAA,EAAA;AAGT,iBAAe,gBAAgB,MAA4C;AACzE,UAAM,EAAE,GAAG,aAAa,QAAQ,CAAC;AAEjC,UAAM,SAA8C;AAEpD,WAAO,MAAM,eAAe;AAAA,MAC1B;AAAA,MACA,OAAO,OAAO;AAAA;AAAA,MACd,SAAS,OAAO;AAAA;AAAA,IAAA,CACjB;AAAA,EAAA;AAGH,iBAAe,iBAEb,MAA6C;AACvC,UAAA,EAAE,GAAG,OAAA,IAAW;AACtB,WAAO,MAAM,gBAAgB;AAAA,MAC3B;AAAA,MACA,MAAM;AAAA,QACJ,GAAG;AAAA,QACH,oBAAoB,OAAO;AAAA,QAC3B,YAAY,OAAO;AAAA,MACrB;AAAA,MACA,OAAO,OAAO;AAAA,MACd,SAAS,OAAO;AAAA,IAAA,CACjB;AAAA,EAAA;AAGY,iBAAA,gCACbA,SACA,QAC6D;;AAC7D,UAAM,kBAA4C,CAAC;AACnD,UAAM,mBAA6C,CAAC;AAEhD,QAAA,CAACA,QAAe,QAAA;AACpB,UAAM,kBACJ,CAAC;AACQ,eAAA,UAAU,OAAO,MAAM;AAChC,UAAI,cAAcA,QAAO,UAAU,WAAW,EAAE;AAAA,QAC9C,OAAO;AAAA,MACT;AACI,UAAA,uBAAuB,QAAS,eAAc,MAAM;AACxD,UAAI,WAAW,aAAa;AAC1B,eAAO,YAAY,YAAY;AAAA,MAAA,OAC1B;AACW,wBAAA,KAAK,GAAG,YAAY,MAAM;AAAA,MAAA;AAG5C,UAAIA,QAAO,YAAY;AACrB,mBAAW,CAAC,YAAY,aAAa,KAAK,OAAO;AAAA,UAC/C,OAAO;AAAA,QAAA,GACN;AACD,gBAAM,yBACJ,CAAC;AACH,qBAAW,gBAAgB,eAAe;AACxC,gBAAI,gBACFA,aAAO,WAAW,UAAU,MAA5BA,mBAAgC,aAAa;AAAA,cAC3C;AAAA;AAEJ,gBAAI,wBAAwB;AAC1B,6BAAe,MAAM;AACnB,gBAAA,gBAAgB,WAAW,cAAc;AAC3C,qCAAuB,KAAK;AAAA,gBAC1B,GAAG,aAAa;AAAA,gBAChB,UAAU,aAAa;AAAA,gBACvB,OAAO,aAAa;AAAA,cAAA,CACrB;AAAA,YAAA,OACI;AACL,+BAAiB,KAAK,IAAI,6CAAc,WAAU,CAAA,CAAG;AAAA,YAAA;AAAA,UACvD;AAGK,iBAAA,WAAW,UAAU,IAAI;AAAA,QAAA;AAAA,MAClC;AAGF,sBAAgB,KAAK,MAAM;AAAA,IAAA;AAE7B,WAAO,OAAO;AAEd,QAAI,gBAAgB,SAAS,KAAK,iBAAiB,SAAS,GAAG;AACrD,cAAA;AAAA,QACN,sDAAsD,MAAM;AAAA,QAC5D;AAAA,UACE;AAAA,UACA;AAAA,QAAA;AAAA,MAEJ;AACM,YAAA,IAAI,MAAM,0BAA0B;AAAA,IAAA;AAGrC,WAAA;AAAA,EAAA;AAGT,iBAAe,eAAe,MAAwC;AACpE,WAAO,MAAM,cAAc;AAAA,MACzB,GAAG;AAAA,MACH;AAAA,IAAA,CACD;AAAA,EAAA;AAGI,SAAA;AAAA,IACL,GAAG;AAAA,IACH,QAAQ,QAAQ;AAAA,IAChB,MAAM;AAAA,IACN;AAAA,IACA,QAAQ;AAAA,IACR,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,QAAQ;AAAA,IACR,MAAM;AAAA,IACN;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,gBAAgB;AAAA,IAChB,iBAAiB;AAAA,IACjB,eAAe;AAAA,EACjB;AACF;"}
|
|
1
|
+
{"version":3,"file":"client.js","sources":["../../src/client.ts"],"sourcesContent":["import type { StandardSchemaV1 } from \"@standard-schema/spec\";\nimport type { Adapter, ExecuteScriptOptions } from \"./adapters/core.js\";\nimport type {\n CreateParams,\n CreateResponse,\n DeleteParams,\n DeleteResponse,\n FieldData,\n FMRecord,\n GenericPortalData,\n GetParams,\n GetResponse,\n GetResponseOne,\n ListParams,\n PortalsWithIds,\n Query,\n UpdateParams,\n UpdateResponse,\n} from \"./client-types.js\";\nimport { FileMakerError } from \"./index.js\";\n\nfunction asNumber(input: string | number): number {\n return typeof input === \"string\" ? Number.parseInt(input, 10) : input;\n}\n\nexport interface ClientObjectProps {\n /**\n * The layout to use by default for all requests. Can be overrridden on each request.\n */\n layout: string;\n schema?: {\n /**\n * The schema for the field data.\n */\n fieldData: StandardSchemaV1<FieldData>;\n /**\n * The schema for the portal data.\n */\n portalData?: Record<string, StandardSchemaV1<FieldData>>;\n };\n}\n\ninterface FetchOptions {\n fetch?: RequestInit;\n}\n\nexport interface IgnoreEmptyResult {\n /**\n * If true, a find that returns no results will retun an empty array instead of throwing an error.\n * @default false\n */\n ignoreEmptyResult?: boolean;\n}\n\nexport interface ContainerUploadArgs<T extends FieldData = FieldData> {\n containerFieldName: keyof T;\n containerFieldRepetition?: string | number;\n file: Blob;\n recordId: number | string;\n modId?: number;\n timeout?: number;\n}\n\nfunction DataApi<\n Fd extends FieldData = FieldData,\n Pd extends GenericPortalData = GenericPortalData,\n Opts extends ClientObjectProps = ClientObjectProps,\n Adp extends Adapter = Adapter,\n>(options: Opts & { adapter: Adp }) {\n type InferredFieldData = Opts[\"schema\"] extends object\n ? StandardSchemaV1.InferOutput<Opts[\"schema\"][\"fieldData\"]>\n : Fd;\n type InferredPortalData = Opts[\"schema\"] extends object\n ? Opts[\"schema\"][\"portalData\"] extends object\n ? {\n [K in keyof Opts[\"schema\"][\"portalData\"]]: StandardSchemaV1.InferOutput<Opts[\"schema\"][\"portalData\"][K]>;\n }\n : Pd\n : Pd;\n\n if (\"zodValidators\" in options) {\n throw new Error(\"zodValidators is no longer supported. Use schema instead, or re-run the typegen command\");\n }\n\n const schema = options.schema;\n const layout = options.layout;\n const {\n create,\n delete: _adapterDelete,\n find,\n get,\n list,\n update,\n layoutMetadata,\n containerUpload,\n executeScript,\n ...otherMethods\n } = options.adapter;\n\n type CreateArgs<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = CreateParams<U> & {\n fieldData: Partial<T>;\n };\n type GetArgs<U extends InferredPortalData = InferredPortalData> = GetParams<U> & {\n recordId: number | string;\n };\n type UpdateArgs<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = UpdateParams<U> & {\n fieldData: Partial<T>;\n recordId: number | string;\n };\n type DeleteArgs = DeleteParams & {\n recordId: number | string;\n };\n type FindArgs<\n T extends FieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n > = ListParams<T, U> & {\n query: Query<T> | Query<T>[];\n timeout?: number;\n };\n\n type ExecuteScriptArgs = Omit<ExecuteScriptOptions, \"layout\">;\n\n /**\n * List all records from a given layout, no find criteria applied.\n */\n async function _list(\n args?: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>>;\n async function _list(\n args?: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const { fetch, timeout, ...params } = args ?? {};\n\n // rename and refactor limit, offset, and sort keys for this request\n if (\"limit\" in params && params.limit !== undefined) {\n Object.assign(params, { _limit: params.limit }).limit = undefined;\n }\n if (\"offset\" in params && params.offset !== undefined) {\n if (params.offset <= 1) {\n params.offset = undefined;\n } else {\n Object.assign(params, { _offset: params.offset }).offset = undefined;\n }\n }\n if (\"sort\" in params && params.sort !== undefined) {\n Object.assign(params, {\n _sort: Array.isArray(params.sort) ? params.sort : [params.sort],\n }).sort = undefined;\n }\n\n const result = await list({\n layout,\n data: params,\n fetch,\n timeout,\n });\n\n if (\n result.dataInfo.foundCount > result.dataInfo.returnedCount &&\n args?.limit === undefined &&\n args?.offset === undefined\n ) {\n // more records found than returned and the user didn't specify a limit or offset, so we should warn them\n console.warn(\n `🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your \"${layout}\" layout. Use the \"listAll\" method to automatically paginate through all records, or specify a \"limit\" and \"offset\" to handle pagination yourself.`,\n );\n }\n\n return await runSchemaValidationAndTransform(schema, result as GetResponse<InferredFieldData, InferredPortalData>);\n }\n\n /**\n * Paginate through all records from a given layout, no find criteria applied.\n * ⚠️ WARNING: Use this method with caution, as it can be slow with large datasets\n */\n async function listAll(\n args?: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<FMRecord<InferredFieldData, InferredPortalData>[]>;\n async function listAll(\n args?: ListParams<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<FMRecord<InferredFieldData, InferredPortalData>[]> {\n let runningData: GetResponse<InferredFieldData, InferredPortalData>[\"data\"] = [];\n const limit = args?.limit ?? 100;\n let offset = args?.offset ?? 1;\n\n while (true) {\n const data = await _list({\n ...args,\n offset,\n });\n runningData = [...runningData, ...data.data];\n if (runningData.length >= data.dataInfo.foundCount) {\n break;\n }\n offset += limit;\n }\n return runningData;\n }\n\n /**\n * Create a new record in a given layout\n */\n async function _create<\n T extends InferredFieldData = InferredFieldData,\n U extends InferredPortalData = InferredPortalData,\n >(args: CreateArgs<T, U> & FetchOptions): Promise<CreateResponse> {\n const { fetch, timeout, ...params } = args ?? {};\n return await create({\n layout,\n data: params,\n fetch,\n timeout,\n });\n }\n\n /**\n * Get a single record by Internal RecordId\n */\n async function _get(\n args: GetArgs<InferredPortalData> & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n\n const result = await get({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n return await runSchemaValidationAndTransform(schema, result as GetResponse<InferredFieldData, InferredPortalData>);\n }\n\n /**\n * Update a single record by internal RecordId\n */\n async function _update(\n args: UpdateArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<UpdateResponse> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n return await update({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n }\n\n /**\n * Delete a single record by internal RecordId\n */\n function deleteRecord(args: DeleteArgs & FetchOptions): Promise<DeleteResponse> {\n args.recordId = asNumber(args.recordId);\n const { recordId, fetch, timeout, ...params } = args;\n\n return _adapterDelete({\n layout,\n data: { ...params, recordId },\n fetch,\n timeout,\n });\n }\n\n /**\n * Find records in a given layout\n */\n async function _find(\n args: FindArgs<InferredFieldData, InferredPortalData> & IgnoreEmptyResult & FetchOptions,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const { query: queryInput, ignoreEmptyResult = false, timeout, fetch, ...params } = args;\n const query = Array.isArray(queryInput) ? queryInput : [queryInput];\n\n // rename and refactor limit, offset, and sort keys for this request\n if (\"offset\" in params && params.offset !== undefined && params.offset <= 1) {\n params.offset = undefined;\n }\n if (\"dateformats\" in params && params.dateformats !== undefined) {\n // reassign dateformats to match FileMaker's expected values\n let dateFormatValue: number;\n if (params.dateformats === \"US\") {\n dateFormatValue = 0;\n } else if (params.dateformats === \"file_locale\") {\n dateFormatValue = 1;\n } else if (params.dateformats === \"ISO8601\") {\n dateFormatValue = 2;\n } else {\n dateFormatValue = 0;\n }\n // @ts-expect-error FM wants a string, so this is fine\n params.dateformats = dateFormatValue.toString();\n }\n const result = (await find({\n data: { ...params, query },\n layout,\n fetch,\n timeout,\n }).catch((e: unknown) => {\n if (ignoreEmptyResult && e instanceof FileMakerError && e.code === \"401\") {\n return { data: [], dataInfo: { foundCount: 0, returnedCount: 0 } };\n }\n throw e;\n })) as GetResponse<InferredFieldData, InferredPortalData>;\n\n if (\n result.dataInfo.foundCount > result.dataInfo.returnedCount &&\n args?.limit === undefined &&\n args?.offset === undefined\n ) {\n // more records found than returned and the user didn't specify a limit or offset\n console.warn(\n `🚨 @proofkit/fmdapi: Loaded only ${result.dataInfo.returnedCount} of the ${result.dataInfo.foundCount} records from your \"${layout}\" layout. Use the \"findAll\" method to automatically paginate through all records, or specify a \"limit\" and \"offset\" to handle pagination yourself.`,\n );\n }\n\n return await runSchemaValidationAndTransform(schema, result);\n }\n\n /**\n * Helper method for `find`. Will only return the first result or throw error if there is more than 1 result.\n */\n async function findOne(\n args: FindArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData>> {\n const result = await _find(args);\n if (result.data.length !== 1) {\n throw new Error(`${result.data.length} records found; expecting exactly 1`);\n }\n const transformedResult = await runSchemaValidationAndTransform(schema, result);\n if (!transformedResult.data[0]) {\n throw new Error(\"No data found\");\n }\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find`. Will only return the first result instead of an array.\n */\n async function findFirst(\n args: FindArgs<InferredFieldData, InferredPortalData> & IgnoreEmptyResult & FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData>> {\n const result = await _find(args);\n const transformedResult = await runSchemaValidationAndTransform(schema, result);\n\n if (!transformedResult.data[0]) {\n throw new Error(\"No data found\");\n }\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find`. Will return the first result or null if no results are found.\n */\n async function maybeFindFirst(\n args: FindArgs<InferredFieldData, InferredPortalData> & IgnoreEmptyResult & FetchOptions,\n ): Promise<GetResponseOne<InferredFieldData, InferredPortalData> | null> {\n const result = await _find({ ...args, ignoreEmptyResult: true });\n const transformedResult = await runSchemaValidationAndTransform(schema, result);\n if (!transformedResult.data[0]) {\n return null;\n }\n return { ...transformedResult, data: transformedResult.data[0] };\n }\n\n /**\n * Helper method for `find` to page through all found results.\n * ⚠️ WARNING: Use with caution as this can be a slow operation with large datasets\n */\n async function findAll(\n args: FindArgs<InferredFieldData, InferredPortalData> & FetchOptions,\n ): Promise<FMRecord<InferredFieldData, InferredPortalData>[]> {\n let runningData: GetResponse<InferredFieldData, InferredPortalData>[\"data\"] = [];\n const limit = args.limit ?? 100;\n let offset = args.offset ?? 1;\n\n while (true) {\n const data = await _find({\n ...args,\n offset,\n ignoreEmptyResult: true,\n });\n runningData = [...runningData, ...data.data];\n if (runningData.length === 0 || runningData.length >= data.dataInfo.foundCount) {\n break;\n }\n offset += limit;\n }\n return runningData;\n }\n\n async function _layoutMetadata(args?: { timeout?: number } & FetchOptions) {\n const { ...restArgs } = args ?? {};\n // Explicitly define the type for params based on FetchOptions\n const params: FetchOptions & { timeout?: number } = restArgs;\n\n return await layoutMetadata({\n layout,\n fetch: params.fetch, // Now should correctly resolve to undefined if not present\n timeout: params.timeout, // Now should correctly resolve to undefined if not present\n });\n }\n\n async function _containerUpload(args: ContainerUploadArgs<InferredFieldData> & FetchOptions) {\n const { ...params } = args;\n return await containerUpload({\n layout,\n data: {\n ...params,\n containerFieldName: params.containerFieldName as string,\n repetition: params.containerFieldRepetition,\n },\n fetch: params.fetch,\n timeout: params.timeout,\n });\n }\n\n async function runSchemaValidationAndTransform(\n schema: ClientObjectProps[\"schema\"],\n result: GetResponse<InferredFieldData, InferredPortalData>,\n ): Promise<GetResponse<InferredFieldData, InferredPortalData>> {\n const fieldDataIssues: StandardSchemaV1.Issue[] = [];\n const portalDataIssues: StandardSchemaV1.Issue[] = [];\n\n if (!schema) {\n return result;\n }\n const transformedData: FMRecord<InferredFieldData, InferredPortalData>[] = [];\n for (const record of result.data) {\n let fieldResult = schema.fieldData[\"~standard\"].validate(record.fieldData);\n if (fieldResult instanceof Promise) {\n fieldResult = await fieldResult;\n }\n if (\"value\" in fieldResult) {\n record.fieldData = fieldResult.value as InferredFieldData;\n } else {\n fieldDataIssues.push(...fieldResult.issues);\n }\n\n if (schema.portalData) {\n for (const [portalName, portalRecords] of Object.entries(record.portalData)) {\n const validatedPortalRecords: PortalsWithIds<GenericPortalData>[] = [];\n for (const portalRecord of portalRecords) {\n let portalResult = schema.portalData[portalName]?.[\"~standard\"].validate(portalRecord);\n if (portalResult instanceof Promise) {\n portalResult = await portalResult;\n }\n if (portalResult && \"value\" in portalResult) {\n validatedPortalRecords.push({\n ...portalResult.value,\n recordId: portalRecord.recordId,\n modId: portalRecord.modId,\n });\n } else {\n portalDataIssues.push(...(portalResult?.issues ?? []));\n }\n }\n // @ts-expect-error We know portalName is a valid key, but can't figure out the right assertions\n record.portalData[portalName] = validatedPortalRecords;\n }\n }\n\n transformedData.push(record);\n }\n result.data = transformedData;\n\n if (fieldDataIssues.length > 0 || portalDataIssues.length > 0) {\n console.error(\n `🚨 @proofkit/fmdapi: Validation issues for layout \"${layout}\". Run the typegen command again to generate the latest field definitions from your layout.`,\n {\n fieldDataIssues,\n portalDataIssues,\n },\n );\n throw new Error(\"Schema validation issues\");\n }\n\n return result;\n }\n\n async function _executeScript(args: ExecuteScriptArgs & FetchOptions) {\n return await executeScript({\n ...args,\n layout,\n });\n }\n\n return {\n ...otherMethods,\n layout: options.layout as Opts[\"layout\"],\n list: _list,\n listAll,\n create: _create,\n get: _get,\n update: _update,\n delete: deleteRecord,\n find: _find,\n findOne,\n findFirst,\n maybeFindFirst,\n findAll,\n layoutMetadata: _layoutMetadata,\n containerUpload: _containerUpload,\n executeScript: _executeScript,\n };\n}\n\nexport default DataApi;\nexport { DataApi };\n"],"names":["schema"],"mappings":";AAqBA,SAAS,SAAS,OAAgC;AAChD,SAAO,OAAO,UAAU,WAAW,OAAO,SAAS,OAAO,EAAE,IAAI;AAClE;AAwCA,SAAS,QAKP,SAAkC;AAYlC,MAAI,mBAAmB,SAAS;AAC9B,UAAM,IAAI,MAAM,yFAAyF;AAAA,EAC3G;AAEA,QAAM,SAAS,QAAQ;AACvB,QAAM,SAAS,QAAQ;AACvB,QAAM;AAAA,IACJ;AAAA,IACA,QAAQ;AAAA,IACR;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,GAAG;AAAA,EAAA,IACD,QAAQ;AAqCZ,iBAAe,MACb,MAC6D;AAC7D,UAAM,EAAE,OAAO,SAAS,GAAG,OAAA,IAAW,QAAQ,CAAA;AAG9C,QAAI,WAAW,UAAU,OAAO,UAAU,QAAW;AACnD,aAAO,OAAO,QAAQ,EAAE,QAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ;AAAA,IAC1D;AACA,QAAI,YAAY,UAAU,OAAO,WAAW,QAAW;AACrD,UAAI,OAAO,UAAU,GAAG;AACtB,eAAO,SAAS;AAAA,MAClB,OAAO;AACL,eAAO,OAAO,QAAQ,EAAE,SAAS,OAAO,OAAA,CAAQ,EAAE,SAAS;AAAA,MAC7D;AAAA,IACF;AACA,QAAI,UAAU,UAAU,OAAO,SAAS,QAAW;AACjD,aAAO,OAAO,QAAQ;AAAA,QACpB,OAAO,MAAM,QAAQ,OAAO,IAAI,IAAI,OAAO,OAAO,CAAC,OAAO,IAAI;AAAA,MAAA,CAC/D,EAAE,OAAO;AAAA,IACZ;AAEA,UAAM,SAAS,MAAM,KAAK;AAAA,MACxB;AAAA,MACA,MAAM;AAAA,MACN;AAAA,MACA;AAAA,IAAA,CACD;AAED,QACE,OAAO,SAAS,aAAa,OAAO,SAAS,kBAC7C,6BAAM,WAAU,WAChB,6BAAM,YAAW,QACjB;AAEA,cAAQ;AAAA,QACN,oCAAoC,OAAO,SAAS,aAAa,WAAW,OAAO,SAAS,UAAU,uBAAuB,MAAM;AAAA,MAAA;AAAA,IAEvI;AAEA,WAAO,MAAM,gCAAgC,QAAQ,MAA4D;AAAA,EACnH;AASA,iBAAe,QACb,MAC4D;AAC5D,QAAI,cAA0E,CAAA;AAC9E,UAAM,SAAQ,6BAAM,UAAS;AAC7B,QAAI,UAAS,6BAAM,WAAU;AAE7B,WAAO,MAAM;AACX,YAAM,OAAO,MAAM,MAAM;AAAA,QACvB,GAAG;AAAA,QACH;AAAA,MAAA,CACD;AACD,oBAAc,CAAC,GAAG,aAAa,GAAG,KAAK,IAAI;AAC3C,UAAI,YAAY,UAAU,KAAK,SAAS,YAAY;AAClD;AAAA,MACF;AACA,gBAAU;AAAA,IACZ;AACA,WAAO;AAAA,EACT;AAKA,iBAAe,QAGb,MAAgE;AAChE,UAAM,EAAE,OAAO,SAAS,GAAG,OAAA,IAAW,QAAQ,CAAA;AAC9C,WAAO,MAAM,OAAO;AAAA,MAClB;AAAA,MACA,MAAM;AAAA,MACN;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EACH;AAKA,iBAAe,KACb,MAC6D;AAC7D,SAAK,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,WAAW;AAEhD,UAAM,SAAS,MAAM,IAAI;AAAA,MACvB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAA;AAAA,MACnB;AAAA,MACA;AAAA,IAAA,CACD;AACD,WAAO,MAAM,gCAAgC,QAAQ,MAA4D;AAAA,EACnH;AAKA,iBAAe,QACb,MACyB;AACzB,SAAK,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,WAAW;AAChD,WAAO,MAAM,OAAO;AAAA,MAClB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAA;AAAA,MACnB;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EACH;AAKA,WAAS,aAAa,MAA0D;AAC9E,SAAK,WAAW,SAAS,KAAK,QAAQ;AACtC,UAAM,EAAE,UAAU,OAAO,SAAS,GAAG,WAAW;AAEhD,WAAO,eAAe;AAAA,MACpB;AAAA,MACA,MAAM,EAAE,GAAG,QAAQ,SAAA;AAAA,MACnB;AAAA,MACA;AAAA,IAAA,CACD;AAAA,EACH;AAKA,iBAAe,MACb,MAC6D;AAC7D,UAAM,EAAE,OAAO,YAAY,oBAAoB,OAAO,SAAS,OAAO,GAAG,OAAA,IAAW;AACpF,UAAM,QAAQ,MAAM,QAAQ,UAAU,IAAI,aAAa,CAAC,UAAU;AAGlE,QAAI,YAAY,UAAU,OAAO,WAAW,UAAa,OAAO,UAAU,GAAG;AAC3E,aAAO,SAAS;AAAA,IAClB;AACA,QAAI,iBAAiB,UAAU,OAAO,gBAAgB,QAAW;AAE/D,UAAI;AACJ,UAAI,OAAO,gBAAgB,MAAM;AAC/B,0BAAkB;AAAA,MACpB,WAAW,OAAO,gBAAgB,eAAe;AAC/C,0BAAkB;AAAA,MACpB,WAAW,OAAO,gBAAgB,WAAW;AAC3C,0BAAkB;AAAA,MACpB,OAAO;AACL,0BAAkB;AAAA,MACpB;AAEA,aAAO,cAAc,gBAAgB,SAAA;AAAA,IACvC;AACA,UAAM,SAAU,MAAM,KAAK;AAAA,MACzB,MAAM,EAAE,GAAG,QAAQ,MAAA;AAAA,MACnB;AAAA,MACA;AAAA,MACA;AAAA,IAAA,CACD,EAAE,MAAM,CAAC,MAAe;AACvB,UAAI,qBAAqB,aAAa,kBAAkB,EAAE,SAAS,OAAO;AACxE,eAAO,EAAE,MAAM,CAAA,GAAI,UAAU,EAAE,YAAY,GAAG,eAAe,IAAE;AAAA,MACjE;AACA,YAAM;AAAA,IACR,CAAC;AAED,QACE,OAAO,SAAS,aAAa,OAAO,SAAS,kBAC7C,6BAAM,WAAU,WAChB,6BAAM,YAAW,QACjB;AAEA,cAAQ;AAAA,QACN,oCAAoC,OAAO,SAAS,aAAa,WAAW,OAAO,SAAS,UAAU,uBAAuB,MAAM;AAAA,MAAA;AAAA,IAEvI;AAEA,WAAO,MAAM,gCAAgC,QAAQ,MAAM;AAAA,EAC7D;AAKA,iBAAe,QACb,MACgE;AAChE,UAAM,SAAS,MAAM,MAAM,IAAI;AAC/B,QAAI,OAAO,KAAK,WAAW,GAAG;AAC5B,YAAM,IAAI,MAAM,GAAG,OAAO,KAAK,MAAM,qCAAqC;AAAA,IAC5E;AACA,UAAM,oBAAoB,MAAM,gCAAgC,QAAQ,MAAM;AAC9E,QAAI,CAAC,kBAAkB,KAAK,CAAC,GAAG;AAC9B,YAAM,IAAI,MAAM,eAAe;AAAA,IACjC;AACA,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAA;AAAA,EAC/D;AAKA,iBAAe,UACb,MACgE;AAChE,UAAM,SAAS,MAAM,MAAM,IAAI;AAC/B,UAAM,oBAAoB,MAAM,gCAAgC,QAAQ,MAAM;AAE9E,QAAI,CAAC,kBAAkB,KAAK,CAAC,GAAG;AAC9B,YAAM,IAAI,MAAM,eAAe;AAAA,IACjC;AACA,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAA;AAAA,EAC/D;AAKA,iBAAe,eACb,MACuE;AACvE,UAAM,SAAS,MAAM,MAAM,EAAE,GAAG,MAAM,mBAAmB,MAAM;AAC/D,UAAM,oBAAoB,MAAM,gCAAgC,QAAQ,MAAM;AAC9E,QAAI,CAAC,kBAAkB,KAAK,CAAC,GAAG;AAC9B,aAAO;AAAA,IACT;AACA,WAAO,EAAE,GAAG,mBAAmB,MAAM,kBAAkB,KAAK,CAAC,EAAA;AAAA,EAC/D;AAMA,iBAAe,QACb,MAC4D;AAC5D,QAAI,cAA0E,CAAA;AAC9E,UAAM,QAAQ,KAAK,SAAS;AAC5B,QAAI,SAAS,KAAK,UAAU;AAE5B,WAAO,MAAM;AACX,YAAM,OAAO,MAAM,MAAM;AAAA,QACvB,GAAG;AAAA,QACH;AAAA,QACA,mBAAmB;AAAA,MAAA,CACpB;AACD,oBAAc,CAAC,GAAG,aAAa,GAAG,KAAK,IAAI;AAC3C,UAAI,YAAY,WAAW,KAAK,YAAY,UAAU,KAAK,SAAS,YAAY;AAC9E;AAAA,MACF;AACA,gBAAU;AAAA,IACZ;AACA,WAAO;AAAA,EACT;AAEA,iBAAe,gBAAgB,MAA4C;AACzE,UAAM,EAAE,GAAG,aAAa,QAAQ,CAAA;AAEhC,UAAM,SAA8C;AAEpD,WAAO,MAAM,eAAe;AAAA,MAC1B;AAAA,MACA,OAAO,OAAO;AAAA;AAAA,MACd,SAAS,OAAO;AAAA;AAAA,IAAA,CACjB;AAAA,EACH;AAEA,iBAAe,iBAAiB,MAA6D;AAC3F,UAAM,EAAE,GAAG,OAAA,IAAW;AACtB,WAAO,MAAM,gBAAgB;AAAA,MAC3B;AAAA,MACA,MAAM;AAAA,QACJ,GAAG;AAAA,QACH,oBAAoB,OAAO;AAAA,QAC3B,YAAY,OAAO;AAAA,MAAA;AAAA,MAErB,OAAO,OAAO;AAAA,MACd,SAAS,OAAO;AAAA,IAAA,CACjB;AAAA,EACH;AAEA,iBAAe,gCACbA,SACA,QAC6D;;AAC7D,UAAM,kBAA4C,CAAA;AAClD,UAAM,mBAA6C,CAAA;AAEnD,QAAI,CAACA,SAAQ;AACX,aAAO;AAAA,IACT;AACA,UAAM,kBAAqE,CAAA;AAC3E,eAAW,UAAU,OAAO,MAAM;AAChC,UAAI,cAAcA,QAAO,UAAU,WAAW,EAAE,SAAS,OAAO,SAAS;AACzE,UAAI,uBAAuB,SAAS;AAClC,sBAAc,MAAM;AAAA,MACtB;AACA,UAAI,WAAW,aAAa;AAC1B,eAAO,YAAY,YAAY;AAAA,MACjC,OAAO;AACL,wBAAgB,KAAK,GAAG,YAAY,MAAM;AAAA,MAC5C;AAEA,UAAIA,QAAO,YAAY;AACrB,mBAAW,CAAC,YAAY,aAAa,KAAK,OAAO,QAAQ,OAAO,UAAU,GAAG;AAC3E,gBAAM,yBAA8D,CAAA;AACpE,qBAAW,gBAAgB,eAAe;AACxC,gBAAI,gBAAeA,aAAO,WAAW,UAAU,MAA5BA,mBAAgC,aAAa,SAAS;AACzE,gBAAI,wBAAwB,SAAS;AACnC,6BAAe,MAAM;AAAA,YACvB;AACA,gBAAI,gBAAgB,WAAW,cAAc;AAC3C,qCAAuB,KAAK;AAAA,gBAC1B,GAAG,aAAa;AAAA,gBAChB,UAAU,aAAa;AAAA,gBACvB,OAAO,aAAa;AAAA,cAAA,CACrB;AAAA,YACH,OAAO;AACL,+BAAiB,KAAK,IAAI,6CAAc,WAAU,CAAA,CAAG;AAAA,YACvD;AAAA,UACF;AAEA,iBAAO,WAAW,UAAU,IAAI;AAAA,QAClC;AAAA,MACF;AAEA,sBAAgB,KAAK,MAAM;AAAA,IAC7B;AACA,WAAO,OAAO;AAEd,QAAI,gBAAgB,SAAS,KAAK,iBAAiB,SAAS,GAAG;AAC7D,cAAQ;AAAA,QACN,sDAAsD,MAAM;AAAA,QAC5D;AAAA,UACE;AAAA,UACA;AAAA,QAAA;AAAA,MACF;AAEF,YAAM,IAAI,MAAM,0BAA0B;AAAA,IAC5C;AAEA,WAAO;AAAA,EACT;AAEA,iBAAe,eAAe,MAAwC;AACpE,WAAO,MAAM,cAAc;AAAA,MACzB,GAAG;AAAA,MACH;AAAA,IAAA,CACD;AAAA,EACH;AAEA,SAAO;AAAA,IACL,GAAG;AAAA,IACH,QAAQ,QAAQ;AAAA,IAChB,MAAM;AAAA,IACN;AAAA,IACA,QAAQ;AAAA,IACR,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,QAAQ;AAAA,IACR,MAAM;AAAA,IACN;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,gBAAgB;AAAA,IAChB,iBAAiB;AAAA,IACjB,eAAe;AAAA,EAAA;AAEnB;"}
|
package/dist/esm/index.d.ts
CHANGED
|
@@ -1,8 +1,7 @@
|
|
|
1
|
-
import { FileMakerError } from './client-types.js';
|
|
2
|
-
import { DataApi } from './client.js';
|
|
3
|
-
export { DataApi, FileMakerError };
|
|
4
|
-
export * from './utils.js';
|
|
5
|
-
export * as clientTypes from './client-types.js';
|
|
6
1
|
export { FetchAdapter } from './adapters/fetch.js';
|
|
2
|
+
export { FmHttpAdapter, type FmHttpAdapterOptions } from './adapters/fm-http.js';
|
|
7
3
|
export { OttoAdapter, type OttoAPIKey } from './adapters/otto.js';
|
|
8
|
-
export default
|
|
4
|
+
export { DataApi, DataApi as default } from './client.js';
|
|
5
|
+
export * as clientTypes from './client-types.js';
|
|
6
|
+
export { FileMakerError } from './client-types.js';
|
|
7
|
+
export * from './utils.js';
|
package/dist/esm/index.js
CHANGED
|
@@ -1,16 +1,18 @@
|
|
|
1
|
+
import { FetchAdapter } from "./adapters/fetch.js";
|
|
2
|
+
import { FmHttpAdapter } from "./adapters/fm-http.js";
|
|
3
|
+
import { OttoAdapter } from "./adapters/otto.js";
|
|
4
|
+
import { default as default2, default as default3 } from "./client.js";
|
|
1
5
|
import * as clientTypes from "./client-types.js";
|
|
2
6
|
import { FileMakerError } from "./client-types.js";
|
|
3
|
-
import DataApi from "./client.js";
|
|
4
7
|
import { removeFMTableNames } from "./utils.js";
|
|
5
|
-
import { FetchAdapter } from "./adapters/fetch.js";
|
|
6
|
-
import { OttoAdapter } from "./adapters/otto.js";
|
|
7
8
|
export {
|
|
8
|
-
DataApi,
|
|
9
|
+
default2 as DataApi,
|
|
9
10
|
FetchAdapter,
|
|
10
11
|
FileMakerError,
|
|
12
|
+
FmHttpAdapter,
|
|
11
13
|
OttoAdapter,
|
|
12
14
|
clientTypes,
|
|
13
|
-
|
|
15
|
+
default3 as default,
|
|
14
16
|
removeFMTableNames
|
|
15
17
|
};
|
|
16
18
|
//# sourceMappingURL=index.js.map
|
package/dist/esm/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sources":[],"sourcesContent":[],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"index.js","sources":[],"sourcesContent":[],"names":[],"mappings":";;;;;;;"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"memory.js","sources":["../../../src/tokenStore/memory.ts"],"sourcesContent":["import type { TokenStoreDefinitions } from \"./types.js\";\n\nexport function memoryStore(): TokenStoreDefinitions {\n const data: Record<string, string> = {};\n return {\n getToken: (key: string): string | null => {\n try {\n return data[key] ?? null;\n } catch {\n return null;\n }\n },\n clearToken: (key: string) => delete data[key],\n setToken: (key: string, value: string): void => {\n data[key] = value;\n },\n };\n}\n\nexport default memoryStore;\n"],"names":[],"mappings":"AAEO,SAAS,cAAqC;AACnD,QAAM,OAA+B,
|
|
1
|
+
{"version":3,"file":"memory.js","sources":["../../../src/tokenStore/memory.ts"],"sourcesContent":["import type { TokenStoreDefinitions } from \"./types.js\";\n\nexport function memoryStore(): TokenStoreDefinitions {\n const data: Record<string, string> = {};\n return {\n getToken: (key: string): string | null => {\n try {\n return data[key] ?? null;\n } catch {\n return null;\n }\n },\n clearToken: (key: string) => delete data[key],\n setToken: (key: string, value: string): void => {\n data[key] = value;\n },\n };\n}\n\nexport default memoryStore;\n"],"names":[],"mappings":"AAEO,SAAS,cAAqC;AACnD,QAAM,OAA+B,CAAA;AACrC,SAAO;AAAA,IACL,UAAU,CAAC,QAA+B;AACxC,UAAI;AACF,eAAO,KAAK,GAAG,KAAK;AAAA,MACtB,QAAQ;AACN,eAAO;AAAA,MACT;AAAA,IACF;AAAA,IACA,YAAY,CAAC,QAAgB,OAAO,KAAK,GAAG;AAAA,IAC5C,UAAU,CAAC,KAAa,UAAwB;AAC9C,WAAK,GAAG,IAAI;AAAA,IACd;AAAA,EAAA;AAEJ;"}
|
|
@@ -1,8 +1,8 @@
|
|
|
1
1
|
type MaybePromise<T> = Promise<T> | T;
|
|
2
|
-
export
|
|
2
|
+
export interface TokenStoreDefinitions {
|
|
3
3
|
getKey?: () => string;
|
|
4
4
|
getToken: (key: string) => MaybePromise<string | null>;
|
|
5
5
|
setToken: (key: string, value: string) => void;
|
|
6
6
|
clearToken: (key: string) => void;
|
|
7
|
-
}
|
|
7
|
+
}
|
|
8
8
|
export {};
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
import { TokenStoreDefinitions } from './types.js';
|
|
2
1
|
import { RedisConfigNodejs } from '@upstash/redis';
|
|
2
|
+
import { TokenStoreDefinitions } from './types.js';
|
|
3
3
|
export declare function upstashTokenStore(config: RedisConfigNodejs, options?: {
|
|
4
4
|
prefix?: string;
|
|
5
5
|
}): TokenStoreDefinitions;
|
package/dist/esm/utils.d.ts
CHANGED
|
@@ -3,15 +3,15 @@ type TransformedFields<T extends object> = {
|
|
|
3
3
|
[K in keyof T as K extends string ? StripFMTableName<K> : K]: T[K];
|
|
4
4
|
};
|
|
5
5
|
export declare function removeFMTableNames<T extends object>(obj: T): TransformedFields<T>;
|
|
6
|
-
export type InferZodPortals<T extends Record<string,
|
|
6
|
+
export type InferZodPortals<T extends Record<string, unknown>> = {
|
|
7
7
|
[K in keyof T]: T[K] extends {
|
|
8
|
-
_def:
|
|
9
|
-
parse: (...args:
|
|
8
|
+
_def: unknown;
|
|
9
|
+
parse: (...args: unknown[]) => unknown;
|
|
10
10
|
} ? ReturnType<T[K]["parse"]> : T[K] extends {
|
|
11
|
-
_def:
|
|
12
|
-
safeParse: (...args:
|
|
11
|
+
_def: unknown;
|
|
12
|
+
safeParse: (...args: unknown[]) => unknown;
|
|
13
13
|
} ? T[K] extends {
|
|
14
|
-
parse: (...args:
|
|
15
|
-
} ? ReturnType<T[K]["parse"]> :
|
|
14
|
+
parse: (...args: unknown[]) => unknown;
|
|
15
|
+
} ? ReturnType<T[K]["parse"]> : unknown : never;
|
|
16
16
|
};
|
|
17
17
|
export {};
|
package/dist/esm/utils.js
CHANGED
|
@@ -1,10 +1,12 @@
|
|
|
1
1
|
function removeFMTableNames(obj) {
|
|
2
2
|
const newObj = {};
|
|
3
3
|
for (const key in obj) {
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
4
|
+
if (Object.hasOwn(obj, key)) {
|
|
5
|
+
const originalKey = key;
|
|
6
|
+
const value = obj[originalKey];
|
|
7
|
+
const mappedKey = typeof key === "string" && key.includes("::") ? key.split("::")[1] : key;
|
|
8
|
+
newObj[mappedKey] = value;
|
|
9
|
+
}
|
|
8
10
|
}
|
|
9
11
|
return newObj;
|
|
10
12
|
}
|
package/dist/esm/utils.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"utils.js","sources":["../../src/utils.ts"],"sourcesContent":["
|
|
1
|
+
{"version":3,"file":"utils.js","sources":["../../src/utils.ts"],"sourcesContent":["type StripFMTableName<K extends PropertyKey> = K extends `${string}::${infer R}` ? R : K;\n\ntype TransformedFields<T extends object> = {\n [K in keyof T as K extends string ? StripFMTableName<K> : K]: T[K];\n};\n\nexport function removeFMTableNames<T extends object>(obj: T): TransformedFields<T> {\n const newObj = {} as TransformedFields<T>;\n for (const key in obj) {\n if (Object.hasOwn(obj, key)) {\n const originalKey = key as keyof T;\n const value = obj[originalKey];\n const mappedKey = (\n typeof key === \"string\" && key.includes(\"::\") ? key.split(\"::\")[1] : key\n ) as keyof TransformedFields<T>;\n\n // Use a temporary index signature cast to assign without any\n (newObj as unknown as Record<PropertyKey, unknown>)[mappedKey as unknown as PropertyKey] = value as unknown;\n }\n }\n return newObj;\n}\n\nexport type InferZodPortals<T extends Record<string, unknown>> = {\n [K in keyof T]: T[K] extends { _def: unknown; parse: (...args: unknown[]) => unknown }\n ? ReturnType<T[K][\"parse\"]>\n : T[K] extends { _def: unknown; safeParse: (...args: unknown[]) => unknown }\n ? T[K] extends { parse: (...args: unknown[]) => unknown }\n ? ReturnType<T[K][\"parse\"]>\n : unknown\n : never;\n};\n"],"names":[],"mappings":"AAMO,SAAS,mBAAqC,KAA8B;AACjF,QAAM,SAAS,CAAA;AACf,aAAW,OAAO,KAAK;AACrB,QAAI,OAAO,OAAO,KAAK,GAAG,GAAG;AAC3B,YAAM,cAAc;AACpB,YAAM,QAAQ,IAAI,WAAW;AAC7B,YAAM,YACJ,OAAO,QAAQ,YAAY,IAAI,SAAS,IAAI,IAAI,IAAI,MAAM,IAAI,EAAE,CAAC,IAAI;AAItE,aAAmD,SAAmC,IAAI;AAAA,IAC7F;AAAA,EACF;AACA,SAAO;AACT;"}
|
package/package.json
CHANGED
|
@@ -1,8 +1,11 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@proofkit/fmdapi",
|
|
3
|
-
"version": "5.0
|
|
3
|
+
"version": "5.1.0-beta.2",
|
|
4
4
|
"description": "FileMaker Data API client",
|
|
5
|
-
"repository":
|
|
5
|
+
"repository": {
|
|
6
|
+
"type": "git",
|
|
7
|
+
"url": "https://github.com/proofgeist/proofkit"
|
|
8
|
+
},
|
|
6
9
|
"author": "Eric <37158449+eluce2@users.noreply.github.com>",
|
|
7
10
|
"license": "MIT",
|
|
8
11
|
"private": false,
|
|
@@ -37,30 +40,26 @@
|
|
|
37
40
|
"./package.json": "./package.json"
|
|
38
41
|
},
|
|
39
42
|
"dependencies": {
|
|
40
|
-
"@standard-schema/spec": "^1.
|
|
41
|
-
"@tanstack/vite-config": "^0.2.
|
|
43
|
+
"@standard-schema/spec": "^1.1.0",
|
|
44
|
+
"@tanstack/vite-config": "^0.2.1",
|
|
42
45
|
"chalk": "5.4.1",
|
|
43
|
-
"commander": "^14.0.
|
|
44
|
-
"dotenv": "^16.
|
|
45
|
-
"fs-extra": "^11.3.
|
|
46
|
+
"commander": "^14.0.2",
|
|
47
|
+
"dotenv": "^16.6.1",
|
|
48
|
+
"fs-extra": "^11.3.3",
|
|
46
49
|
"ts-morph": "^26.0.0",
|
|
47
|
-
"vite": "^6.
|
|
48
|
-
"zod": "3.
|
|
50
|
+
"vite": "^6.4.1",
|
|
51
|
+
"zod": "^4.3.5"
|
|
49
52
|
},
|
|
50
53
|
"devDependencies": {
|
|
54
|
+
"@tanstack/intent": "^0.0.19",
|
|
51
55
|
"@types/fs-extra": "^11.0.4",
|
|
52
|
-
"@types/node": "^22.
|
|
53
|
-
"@
|
|
54
|
-
"
|
|
55
|
-
"
|
|
56
|
-
"eslint": "^9.23.0",
|
|
57
|
-
"eslint-plugin-react": "^7.37.4",
|
|
58
|
-
"knip": "^5.56.0",
|
|
59
|
-
"prettier": "^3.5.3",
|
|
60
|
-
"publint": "^0.3.12",
|
|
56
|
+
"@types/node": "^22.19.5",
|
|
57
|
+
"@upstash/redis": "^1.36.1",
|
|
58
|
+
"knip": "^5.80.2",
|
|
59
|
+
"publint": "^0.3.16",
|
|
61
60
|
"ts-toolbelt": "^9.6.0",
|
|
62
|
-
"typescript": "^5.9.
|
|
63
|
-
"vitest": "^
|
|
61
|
+
"typescript": "^5.9.3",
|
|
62
|
+
"vitest": "^4.0.17"
|
|
64
63
|
},
|
|
65
64
|
"engines": {
|
|
66
65
|
"node": ">=18.0.0"
|
|
@@ -71,7 +70,10 @@
|
|
|
71
70
|
"dist-browser",
|
|
72
71
|
"tokenStore",
|
|
73
72
|
"utils",
|
|
74
|
-
"stubs"
|
|
73
|
+
"stubs",
|
|
74
|
+
"skills",
|
|
75
|
+
"bin",
|
|
76
|
+
"!skills/_artifacts"
|
|
75
77
|
],
|
|
76
78
|
"keywords": [
|
|
77
79
|
"filemaker",
|
|
@@ -82,18 +84,27 @@
|
|
|
82
84
|
"fmrest",
|
|
83
85
|
"fmdapi",
|
|
84
86
|
"proofgeist",
|
|
85
|
-
"fm-dapi"
|
|
87
|
+
"fm-dapi",
|
|
88
|
+
"tanstack-intent"
|
|
86
89
|
],
|
|
90
|
+
"bin": {
|
|
91
|
+
"intent": "./bin/intent.js"
|
|
92
|
+
},
|
|
87
93
|
"scripts": {
|
|
88
94
|
"build": "tsc && vite build && publint --strict",
|
|
89
95
|
"build:watch": "tsc && vite build --watch",
|
|
90
|
-
"check-format": "
|
|
91
|
-
"format": "
|
|
96
|
+
"check-format": "biome format --check .",
|
|
97
|
+
"format": "biome format --write .",
|
|
92
98
|
"dev": "tsc --watch",
|
|
93
99
|
"ci": "pnpm build && pnpm check-format && pnpm publint --strict && pnpm test",
|
|
94
|
-
"test": "
|
|
100
|
+
"test": "vitest run",
|
|
101
|
+
"test:e2e": "doppler run -- vitest run tests/e2e",
|
|
102
|
+
"capture": "doppler run -- npx tsx scripts/capture-responses.ts",
|
|
103
|
+
"typecheck": "tsc --noEmit",
|
|
95
104
|
"changeset": "changeset",
|
|
96
105
|
"release": "pnpm build && changeset publish --access public",
|
|
97
|
-
"knip": "knip"
|
|
106
|
+
"knip": "knip",
|
|
107
|
+
"lint": "biome check . --write",
|
|
108
|
+
"lint:summary": "biome check . --reporter=summary"
|
|
98
109
|
}
|
|
99
110
|
}
|