@credal/actions 0.1.48 → 0.1.50
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/actions/actionMapper.js +6 -0
- package/dist/actions/autogen/templates.d.ts +1 -0
- package/dist/actions/autogen/templates.js +62 -2
- package/dist/actions/autogen/types.d.ts +49 -0
- package/dist/actions/autogen/types.js +18 -2
- package/dist/actions/groups.js +5 -1
- package/dist/actions/providers/confluence/updatePage.js +15 -14
- package/dist/actions/providers/firecrawl/deepResearch.d.ts +3 -0
- package/dist/actions/providers/{generic/genericApiCall.js → firecrawl/deepResearch.js} +21 -21
- package/dist/actions/providers/salesforce/getSalesforceRecordsByQuery.js +5 -1
- package/dist/actions/providers/workday/requestTimeOff.d.ts +23 -0
- package/dist/actions/providers/workday/requestTimeOff.js +88 -0
- package/package.json +1 -1
- package/dist/actions/providers/generic/fillTemplateAction.d.ts +0 -7
- package/dist/actions/providers/generic/fillTemplateAction.js +0 -18
- package/dist/actions/providers/generic/genericApiCall.d.ts +0 -3
- package/dist/actions/providers/googlemaps/nearbysearch.d.ts +0 -3
- package/dist/actions/providers/googlemaps/nearbysearch.js +0 -96
- package/dist/actions/providers/snowflake/runSnowflakeQueryWriteResultsToS3.d.ts +0 -3
- package/dist/actions/providers/snowflake/runSnowflakeQueryWriteResultsToS3.js +0 -154
- package/dist/actions/providers/x/scrapeTweetDataWithNitter.d.ts +0 -3
- package/dist/actions/providers/x/scrapeTweetDataWithNitter.js +0 -45
@@ -73,6 +73,7 @@ const updateDocument_1 = __importDefault(require("./providers/microsoft/updateDo
|
|
73
73
|
const createDocument_1 = __importDefault(require("./providers/microsoft/createDocument"));
|
74
74
|
const getDocument_1 = __importDefault(require("./providers/microsoft/getDocument"));
|
75
75
|
const fetchSalesforceSchema_1 = __importDefault(require("./providers/salesforce/fetchSalesforceSchema"));
|
76
|
+
const deepResearch_1 = __importDefault(require("./providers/firecrawl/deepResearch"));
|
76
77
|
exports.ActionMapper = {
|
77
78
|
generic: {
|
78
79
|
fillTemplate: {
|
@@ -268,6 +269,11 @@ exports.ActionMapper = {
|
|
268
269
|
paramsSchema: types_1.firecrawlScrapeTweetDataWithNitterParamsSchema,
|
269
270
|
outputSchema: types_1.firecrawlScrapeTweetDataWithNitterOutputSchema,
|
270
271
|
},
|
272
|
+
deepResearch: {
|
273
|
+
fn: deepResearch_1.default,
|
274
|
+
paramsSchema: types_1.firecrawlScrapeUrlParamsSchema,
|
275
|
+
outputSchema: types_1.firecrawlScrapeUrlOutputSchema,
|
276
|
+
},
|
271
277
|
},
|
272
278
|
resend: {
|
273
279
|
sendEmail: {
|
@@ -31,6 +31,7 @@ export declare const snowflakeGetRowByFieldValueDefinition: ActionTemplate;
|
|
31
31
|
export declare const snowflakeRunSnowflakeQueryDefinition: ActionTemplate;
|
32
32
|
export declare const openstreetmapGetLatitudeLongitudeFromLocationDefinition: ActionTemplate;
|
33
33
|
export declare const nwsGetForecastForLocationDefinition: ActionTemplate;
|
34
|
+
export declare const firecrawlDeepResearchDefinition: ActionTemplate;
|
34
35
|
export declare const firecrawlScrapeUrlDefinition: ActionTemplate;
|
35
36
|
export declare const firecrawlScrapeTweetDataWithNitterDefinition: ActionTemplate;
|
36
37
|
export declare const resendSendEmailDefinition: ActionTemplate;
|
@@ -1,7 +1,7 @@
|
|
1
1
|
"use strict";
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
3
|
-
exports.
|
4
|
-
exports.githubCreatePullRequestDefinition = exports.githubCreateBranchDefinition = exports.githubCreateOrUpdateFileDefinition = exports.microsoftGetDocumentDefinition = exports.microsoftMessageTeamsChannelDefinition = exports.microsoftMessageTeamsChatDefinition = exports.microsoftUpdateSpreadsheetDefinition = exports.microsoftUpdateDocumentDefinition = exports.microsoftCreateDocumentDefinition = exports.salesforceFetchSalesforceSchemaByObjectDefinition = exports.salesforceGetRecordDefinition = exports.salesforceGetSalesforceRecordsByQueryDefinition = exports.salesforceGenerateSalesReportDefinition = exports.salesforceCreateCaseDefinition = exports.salesforceUpdateRecordDefinition = exports.ashbyUpdateCandidateDefinition = exports.ashbyCreateCandidateDefinition = exports.ashbyListCandidateNotesDefinition = void 0;
|
3
|
+
exports.ashbyListCandidatesDefinition = exports.ashbyAddCandidateToProjectDefinition = exports.ashbyGetCandidateInfoDefinition = exports.ashbyCreateNoteDefinition = exports.lookerEnableUserByEmailDefinition = exports.finnhubGetBasicFinancialsDefinition = exports.finnhubSymbolLookupDefinition = exports.googleOauthUpdatePresentationDefinition = exports.googleOauthCreatePresentationDefinition = exports.googleOauthUpdateSpreadsheetDefinition = exports.googleOauthCreateSpreadsheetDefinition = exports.googleOauthScheduleCalendarMeetingDefinition = exports.googleOauthUpdateDocDefinition = exports.googleOauthCreateNewGoogleDocDefinition = exports.resendSendEmailDefinition = exports.firecrawlScrapeTweetDataWithNitterDefinition = exports.firecrawlScrapeUrlDefinition = exports.firecrawlDeepResearchDefinition = exports.nwsGetForecastForLocationDefinition = exports.openstreetmapGetLatitudeLongitudeFromLocationDefinition = exports.snowflakeRunSnowflakeQueryDefinition = exports.snowflakeGetRowByFieldValueDefinition = exports.mongoInsertMongoDocDefinition = exports.xCreateShareXPostUrlDefinition = exports.linkedinCreateShareLinkedinPostUrlDefinition = exports.zendeskAssignTicketDefinition = exports.zendeskAddCommentToTicketDefinition = exports.zendeskUpdateTicketStatusDefinition = exports.zendeskGetTicketDetailsDefinition = exports.zendeskCreateZendeskTicketDefinition = exports.credalCallCopilotDefinition = exports.googlemapsNearbysearchRestaurantsDefinition = exports.googlemapsValidateAddressDefinition = exports.jiraUpdateJiraTicketStatusDefinition = exports.jiraUpdateJiraTicketDetailsDefinition = exports.jiraGetJiraTicketHistoryDefinition = exports.jiraGetJiraTicketDetailsDefinition = exports.jiraCreateJiraTicketDefinition = exports.jiraCommentJiraTicketDefinition = exports.jiraAssignJiraTicketDefinition = exports.confluenceFetchPageContentDefinition = exports.confluenceOverwritePageDefinition = exports.mathAddDefinition = exports.slackGetChannelMessagesDefinition = exports.slackListConversationsDefinition = exports.slackSendMessageDefinition = exports.asanaUpdateTaskDefinition = exports.asanaCreateTaskDefinition = exports.asanaCommentTaskDefinition = exports.genericFillTemplateDefinition = void 0;
|
4
|
+
exports.githubCreatePullRequestDefinition = exports.githubCreateBranchDefinition = exports.githubCreateOrUpdateFileDefinition = exports.microsoftGetDocumentDefinition = exports.microsoftMessageTeamsChannelDefinition = exports.microsoftMessageTeamsChatDefinition = exports.microsoftUpdateSpreadsheetDefinition = exports.microsoftUpdateDocumentDefinition = exports.microsoftCreateDocumentDefinition = exports.salesforceFetchSalesforceSchemaByObjectDefinition = exports.salesforceGetRecordDefinition = exports.salesforceGetSalesforceRecordsByQueryDefinition = exports.salesforceGenerateSalesReportDefinition = exports.salesforceCreateCaseDefinition = exports.salesforceUpdateRecordDefinition = exports.ashbyUpdateCandidateDefinition = exports.ashbyCreateCandidateDefinition = exports.ashbyListCandidateNotesDefinition = exports.ashbySearchCandidatesDefinition = void 0;
|
5
5
|
exports.genericFillTemplateDefinition = {
|
6
6
|
description: "Simple utility that takes a template and returns it filled in",
|
7
7
|
scopes: [],
|
@@ -1519,6 +1519,66 @@ exports.nwsGetForecastForLocationDefinition = {
|
|
1519
1519
|
name: "getForecastForLocation",
|
1520
1520
|
provider: "nws",
|
1521
1521
|
};
|
1522
|
+
exports.firecrawlDeepResearchDefinition = {
|
1523
|
+
description: "Deep research on a topic using Firecrawl",
|
1524
|
+
scopes: [],
|
1525
|
+
parameters: {
|
1526
|
+
type: "object",
|
1527
|
+
required: ["query"],
|
1528
|
+
properties: {
|
1529
|
+
query: {
|
1530
|
+
type: "string",
|
1531
|
+
description: "The query to search for",
|
1532
|
+
},
|
1533
|
+
maxDepth: {
|
1534
|
+
type: "number",
|
1535
|
+
description: "The maximum depth of the search",
|
1536
|
+
},
|
1537
|
+
timeLimit: {
|
1538
|
+
type: "number",
|
1539
|
+
description: "The time limit for the search in seconds",
|
1540
|
+
},
|
1541
|
+
maxUrls: {
|
1542
|
+
type: "number",
|
1543
|
+
description: "The maximum number of URLs to scrape",
|
1544
|
+
},
|
1545
|
+
},
|
1546
|
+
},
|
1547
|
+
output: {
|
1548
|
+
type: "object",
|
1549
|
+
required: ["finalAnalysis", "sources"],
|
1550
|
+
properties: {
|
1551
|
+
finalAnalysis: {
|
1552
|
+
type: "string",
|
1553
|
+
description: "The content of the research",
|
1554
|
+
},
|
1555
|
+
sources: {
|
1556
|
+
type: "array",
|
1557
|
+
description: "The sources of the research",
|
1558
|
+
items: {
|
1559
|
+
type: "object",
|
1560
|
+
required: ["url", "title"],
|
1561
|
+
properties: {
|
1562
|
+
url: {
|
1563
|
+
type: "string",
|
1564
|
+
description: "The URL of the source",
|
1565
|
+
},
|
1566
|
+
title: {
|
1567
|
+
type: "string",
|
1568
|
+
description: "The title of the source",
|
1569
|
+
},
|
1570
|
+
description: {
|
1571
|
+
type: "string",
|
1572
|
+
description: "The description of the source",
|
1573
|
+
},
|
1574
|
+
},
|
1575
|
+
},
|
1576
|
+
},
|
1577
|
+
},
|
1578
|
+
},
|
1579
|
+
name: "deepResearch",
|
1580
|
+
provider: "firecrawl",
|
1581
|
+
};
|
1522
1582
|
exports.firecrawlScrapeUrlDefinition = {
|
1523
1583
|
description: "Scrape a URL and get website content using Firecrawl",
|
1524
1584
|
scopes: [],
|
@@ -1337,6 +1337,55 @@ export declare const nwsGetForecastForLocationOutputSchema: z.ZodObject<{
|
|
1337
1337
|
}>;
|
1338
1338
|
export type nwsGetForecastForLocationOutputType = z.infer<typeof nwsGetForecastForLocationOutputSchema>;
|
1339
1339
|
export type nwsGetForecastForLocationFunction = ActionFunction<nwsGetForecastForLocationParamsType, AuthParamsType, nwsGetForecastForLocationOutputType>;
|
1340
|
+
export declare const firecrawlDeepResearchParamsSchema: z.ZodObject<{
|
1341
|
+
query: z.ZodString;
|
1342
|
+
maxDepth: z.ZodOptional<z.ZodNumber>;
|
1343
|
+
timeLimit: z.ZodOptional<z.ZodNumber>;
|
1344
|
+
maxUrls: z.ZodOptional<z.ZodNumber>;
|
1345
|
+
}, "strip", z.ZodTypeAny, {
|
1346
|
+
query: string;
|
1347
|
+
maxDepth?: number | undefined;
|
1348
|
+
timeLimit?: number | undefined;
|
1349
|
+
maxUrls?: number | undefined;
|
1350
|
+
}, {
|
1351
|
+
query: string;
|
1352
|
+
maxDepth?: number | undefined;
|
1353
|
+
timeLimit?: number | undefined;
|
1354
|
+
maxUrls?: number | undefined;
|
1355
|
+
}>;
|
1356
|
+
export type firecrawlDeepResearchParamsType = z.infer<typeof firecrawlDeepResearchParamsSchema>;
|
1357
|
+
export declare const firecrawlDeepResearchOutputSchema: z.ZodObject<{
|
1358
|
+
finalAnalysis: z.ZodString;
|
1359
|
+
sources: z.ZodArray<z.ZodObject<{
|
1360
|
+
url: z.ZodString;
|
1361
|
+
title: z.ZodString;
|
1362
|
+
description: z.ZodOptional<z.ZodString>;
|
1363
|
+
}, "strip", z.ZodTypeAny, {
|
1364
|
+
title: string;
|
1365
|
+
url: string;
|
1366
|
+
description?: string | undefined;
|
1367
|
+
}, {
|
1368
|
+
title: string;
|
1369
|
+
url: string;
|
1370
|
+
description?: string | undefined;
|
1371
|
+
}>, "many">;
|
1372
|
+
}, "strip", z.ZodTypeAny, {
|
1373
|
+
finalAnalysis: string;
|
1374
|
+
sources: {
|
1375
|
+
title: string;
|
1376
|
+
url: string;
|
1377
|
+
description?: string | undefined;
|
1378
|
+
}[];
|
1379
|
+
}, {
|
1380
|
+
finalAnalysis: string;
|
1381
|
+
sources: {
|
1382
|
+
title: string;
|
1383
|
+
url: string;
|
1384
|
+
description?: string | undefined;
|
1385
|
+
}[];
|
1386
|
+
}>;
|
1387
|
+
export type firecrawlDeepResearchOutputType = z.infer<typeof firecrawlDeepResearchOutputSchema>;
|
1388
|
+
export type firecrawlDeepResearchFunction = ActionFunction<firecrawlDeepResearchParamsType, AuthParamsType, firecrawlDeepResearchOutputType>;
|
1340
1389
|
export declare const firecrawlScrapeUrlParamsSchema: z.ZodObject<{
|
1341
1390
|
url: z.ZodString;
|
1342
1391
|
}, "strip", z.ZodTypeAny, {
|
@@ -1,8 +1,8 @@
|
|
1
1
|
"use strict";
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
3
3
|
exports.zendeskAssignTicketParamsSchema = exports.zendeskAddCommentToTicketOutputSchema = exports.zendeskAddCommentToTicketParamsSchema = exports.zendeskUpdateTicketStatusOutputSchema = exports.zendeskUpdateTicketStatusParamsSchema = exports.zendeskGetTicketDetailsOutputSchema = exports.zendeskGetTicketDetailsParamsSchema = exports.zendeskCreateZendeskTicketOutputSchema = exports.zendeskCreateZendeskTicketParamsSchema = exports.credalCallCopilotOutputSchema = exports.credalCallCopilotParamsSchema = exports.googlemapsNearbysearchRestaurantsOutputSchema = exports.googlemapsNearbysearchRestaurantsParamsSchema = exports.googlemapsValidateAddressOutputSchema = exports.googlemapsValidateAddressParamsSchema = exports.jiraUpdateJiraTicketStatusOutputSchema = exports.jiraUpdateJiraTicketStatusParamsSchema = exports.jiraUpdateJiraTicketDetailsOutputSchema = exports.jiraUpdateJiraTicketDetailsParamsSchema = exports.jiraGetJiraTicketHistoryOutputSchema = exports.jiraGetJiraTicketHistoryParamsSchema = exports.jiraGetJiraTicketDetailsOutputSchema = exports.jiraGetJiraTicketDetailsParamsSchema = exports.jiraCreateJiraTicketOutputSchema = exports.jiraCreateJiraTicketParamsSchema = exports.jiraCommentJiraTicketOutputSchema = exports.jiraCommentJiraTicketParamsSchema = exports.jiraAssignJiraTicketOutputSchema = exports.jiraAssignJiraTicketParamsSchema = exports.confluenceFetchPageContentOutputSchema = exports.confluenceFetchPageContentParamsSchema = exports.confluenceOverwritePageOutputSchema = exports.confluenceOverwritePageParamsSchema = exports.mathAddOutputSchema = exports.mathAddParamsSchema = exports.slackGetChannelMessagesOutputSchema = exports.slackGetChannelMessagesParamsSchema = exports.slackListConversationsOutputSchema = exports.slackListConversationsParamsSchema = exports.slackSendMessageOutputSchema = exports.slackSendMessageParamsSchema = exports.asanaUpdateTaskOutputSchema = exports.asanaUpdateTaskParamsSchema = exports.asanaCreateTaskOutputSchema = exports.asanaCreateTaskParamsSchema = exports.asanaCommentTaskOutputSchema = exports.asanaCommentTaskParamsSchema = exports.genericFillTemplateOutputSchema = exports.genericFillTemplateParamsSchema = exports.AuthParamsSchema = void 0;
|
4
|
-
exports.
|
5
|
-
exports.githubCreatePullRequestOutputSchema = exports.githubCreatePullRequestParamsSchema = exports.githubCreateBranchOutputSchema = exports.githubCreateBranchParamsSchema = exports.githubCreateOrUpdateFileOutputSchema = exports.githubCreateOrUpdateFileParamsSchema = exports.microsoftGetDocumentOutputSchema = exports.microsoftGetDocumentParamsSchema = exports.microsoftMessageTeamsChannelOutputSchema = exports.microsoftMessageTeamsChannelParamsSchema = exports.microsoftMessageTeamsChatOutputSchema = exports.microsoftMessageTeamsChatParamsSchema = exports.microsoftUpdateSpreadsheetOutputSchema = exports.microsoftUpdateSpreadsheetParamsSchema = exports.microsoftUpdateDocumentOutputSchema = exports.microsoftUpdateDocumentParamsSchema = exports.microsoftCreateDocumentOutputSchema = exports.microsoftCreateDocumentParamsSchema = exports.salesforceFetchSalesforceSchemaByObjectOutputSchema = exports.salesforceFetchSalesforceSchemaByObjectParamsSchema = exports.salesforceGetRecordOutputSchema = exports.salesforceGetRecordParamsSchema = exports.salesforceGetSalesforceRecordsByQueryOutputSchema = exports.salesforceGetSalesforceRecordsByQueryParamsSchema = exports.salesforceGenerateSalesReportOutputSchema = exports.salesforceGenerateSalesReportParamsSchema = exports.salesforceCreateCaseOutputSchema = exports.salesforceCreateCaseParamsSchema = exports.salesforceUpdateRecordOutputSchema = exports.salesforceUpdateRecordParamsSchema = exports.ashbyUpdateCandidateOutputSchema = exports.ashbyUpdateCandidateParamsSchema = exports.ashbyCreateCandidateOutputSchema = exports.ashbyCreateCandidateParamsSchema = exports.ashbyListCandidateNotesOutputSchema = exports.ashbyListCandidateNotesParamsSchema = exports.ashbySearchCandidatesOutputSchema = void 0;
|
4
|
+
exports.ashbyListCandidatesParamsSchema = exports.ashbyAddCandidateToProjectOutputSchema = exports.ashbyAddCandidateToProjectParamsSchema = exports.ashbyGetCandidateInfoOutputSchema = exports.ashbyGetCandidateInfoParamsSchema = exports.ashbyCreateNoteOutputSchema = exports.ashbyCreateNoteParamsSchema = exports.lookerEnableUserByEmailOutputSchema = exports.lookerEnableUserByEmailParamsSchema = exports.finnhubGetBasicFinancialsOutputSchema = exports.finnhubGetBasicFinancialsParamsSchema = exports.finnhubSymbolLookupOutputSchema = exports.finnhubSymbolLookupParamsSchema = exports.googleOauthUpdatePresentationOutputSchema = exports.googleOauthUpdatePresentationParamsSchema = exports.googleOauthCreatePresentationOutputSchema = exports.googleOauthCreatePresentationParamsSchema = exports.googleOauthUpdateSpreadsheetOutputSchema = exports.googleOauthUpdateSpreadsheetParamsSchema = exports.googleOauthCreateSpreadsheetOutputSchema = exports.googleOauthCreateSpreadsheetParamsSchema = exports.googleOauthScheduleCalendarMeetingOutputSchema = exports.googleOauthScheduleCalendarMeetingParamsSchema = exports.googleOauthUpdateDocOutputSchema = exports.googleOauthUpdateDocParamsSchema = exports.googleOauthCreateNewGoogleDocOutputSchema = exports.googleOauthCreateNewGoogleDocParamsSchema = exports.resendSendEmailOutputSchema = exports.resendSendEmailParamsSchema = exports.firecrawlScrapeTweetDataWithNitterOutputSchema = exports.firecrawlScrapeTweetDataWithNitterParamsSchema = exports.firecrawlScrapeUrlOutputSchema = exports.firecrawlScrapeUrlParamsSchema = exports.firecrawlDeepResearchOutputSchema = exports.firecrawlDeepResearchParamsSchema = exports.nwsGetForecastForLocationOutputSchema = exports.nwsGetForecastForLocationParamsSchema = exports.openstreetmapGetLatitudeLongitudeFromLocationOutputSchema = exports.openstreetmapGetLatitudeLongitudeFromLocationParamsSchema = exports.snowflakeRunSnowflakeQueryOutputSchema = exports.snowflakeRunSnowflakeQueryParamsSchema = exports.snowflakeGetRowByFieldValueOutputSchema = exports.snowflakeGetRowByFieldValueParamsSchema = exports.mongoInsertMongoDocOutputSchema = exports.mongoInsertMongoDocParamsSchema = exports.xCreateShareXPostUrlOutputSchema = exports.xCreateShareXPostUrlParamsSchema = exports.linkedinCreateShareLinkedinPostUrlOutputSchema = exports.linkedinCreateShareLinkedinPostUrlParamsSchema = exports.zendeskAssignTicketOutputSchema = void 0;
|
5
|
+
exports.githubCreatePullRequestOutputSchema = exports.githubCreatePullRequestParamsSchema = exports.githubCreateBranchOutputSchema = exports.githubCreateBranchParamsSchema = exports.githubCreateOrUpdateFileOutputSchema = exports.githubCreateOrUpdateFileParamsSchema = exports.microsoftGetDocumentOutputSchema = exports.microsoftGetDocumentParamsSchema = exports.microsoftMessageTeamsChannelOutputSchema = exports.microsoftMessageTeamsChannelParamsSchema = exports.microsoftMessageTeamsChatOutputSchema = exports.microsoftMessageTeamsChatParamsSchema = exports.microsoftUpdateSpreadsheetOutputSchema = exports.microsoftUpdateSpreadsheetParamsSchema = exports.microsoftUpdateDocumentOutputSchema = exports.microsoftUpdateDocumentParamsSchema = exports.microsoftCreateDocumentOutputSchema = exports.microsoftCreateDocumentParamsSchema = exports.salesforceFetchSalesforceSchemaByObjectOutputSchema = exports.salesforceFetchSalesforceSchemaByObjectParamsSchema = exports.salesforceGetRecordOutputSchema = exports.salesforceGetRecordParamsSchema = exports.salesforceGetSalesforceRecordsByQueryOutputSchema = exports.salesforceGetSalesforceRecordsByQueryParamsSchema = exports.salesforceGenerateSalesReportOutputSchema = exports.salesforceGenerateSalesReportParamsSchema = exports.salesforceCreateCaseOutputSchema = exports.salesforceCreateCaseParamsSchema = exports.salesforceUpdateRecordOutputSchema = exports.salesforceUpdateRecordParamsSchema = exports.ashbyUpdateCandidateOutputSchema = exports.ashbyUpdateCandidateParamsSchema = exports.ashbyCreateCandidateOutputSchema = exports.ashbyCreateCandidateParamsSchema = exports.ashbyListCandidateNotesOutputSchema = exports.ashbyListCandidateNotesParamsSchema = exports.ashbySearchCandidatesOutputSchema = exports.ashbySearchCandidatesParamsSchema = exports.ashbyListCandidatesOutputSchema = void 0;
|
6
6
|
const zod_1 = require("zod");
|
7
7
|
exports.AuthParamsSchema = zod_1.z.object({
|
8
8
|
authToken: zod_1.z.string().optional(),
|
@@ -459,6 +459,22 @@ exports.nwsGetForecastForLocationOutputSchema = zod_1.z.object({
|
|
459
459
|
})
|
460
460
|
.optional(),
|
461
461
|
});
|
462
|
+
exports.firecrawlDeepResearchParamsSchema = zod_1.z.object({
|
463
|
+
query: zod_1.z.string().describe("The query to search for"),
|
464
|
+
maxDepth: zod_1.z.number().describe("The maximum depth of the search").optional(),
|
465
|
+
timeLimit: zod_1.z.number().describe("The time limit for the search in seconds").optional(),
|
466
|
+
maxUrls: zod_1.z.number().describe("The maximum number of URLs to scrape").optional(),
|
467
|
+
});
|
468
|
+
exports.firecrawlDeepResearchOutputSchema = zod_1.z.object({
|
469
|
+
finalAnalysis: zod_1.z.string().describe("The content of the research"),
|
470
|
+
sources: zod_1.z
|
471
|
+
.array(zod_1.z.object({
|
472
|
+
url: zod_1.z.string().describe("The URL of the source"),
|
473
|
+
title: zod_1.z.string().describe("The title of the source"),
|
474
|
+
description: zod_1.z.string().describe("The description of the source").optional(),
|
475
|
+
}))
|
476
|
+
.describe("The sources of the research"),
|
477
|
+
});
|
462
478
|
exports.firecrawlScrapeUrlParamsSchema = zod_1.z.object({ url: zod_1.z.string().describe("The URL to scrape") });
|
463
479
|
exports.firecrawlScrapeUrlOutputSchema = zod_1.z.object({ content: zod_1.z.string().describe("The content of the URL") });
|
464
480
|
exports.firecrawlScrapeTweetDataWithNitterParamsSchema = zod_1.z.object({
|
package/dist/actions/groups.js
CHANGED
@@ -84,7 +84,11 @@ exports.ACTION_GROUPS = {
|
|
84
84
|
},
|
85
85
|
FIRECRAWL: {
|
86
86
|
description: "Actions for interacting with Firecrawl",
|
87
|
-
actions: [
|
87
|
+
actions: [
|
88
|
+
templates_1.firecrawlScrapeUrlDefinition,
|
89
|
+
templates_1.firecrawlScrapeTweetDataWithNitterDefinition,
|
90
|
+
templates_1.firecrawlDeepResearchDefinition,
|
91
|
+
],
|
88
92
|
},
|
89
93
|
RESEND: {
|
90
94
|
description: "Action for sending an email",
|
@@ -8,28 +8,30 @@ var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, ge
|
|
8
8
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
9
|
});
|
10
10
|
};
|
11
|
+
var __importDefault = (this && this.__importDefault) || function (mod) {
|
12
|
+
return (mod && mod.__esModule) ? mod : { "default": mod };
|
13
|
+
};
|
11
14
|
Object.defineProperty(exports, "__esModule", { value: true });
|
12
|
-
const
|
13
|
-
function
|
14
|
-
|
15
|
+
const axios_1 = __importDefault(require("axios"));
|
16
|
+
function getConfluenceApi(baseUrl, username, apiToken) {
|
17
|
+
const api = axios_1.default.create({
|
15
18
|
baseURL: baseUrl,
|
16
19
|
headers: {
|
17
20
|
Accept: "application/json",
|
21
|
+
// Tokens are associated with a specific user.
|
18
22
|
Authorization: `Basic ${Buffer.from(`${username}:${apiToken}`).toString("base64")}`,
|
19
23
|
},
|
20
|
-
};
|
24
|
+
});
|
25
|
+
return api;
|
21
26
|
}
|
22
27
|
const confluenceUpdatePage = (_a) => __awaiter(void 0, [_a], void 0, function* ({ params, authParams, }) {
|
23
|
-
const { pageId, content, title } = params;
|
24
|
-
const { baseUrl, authToken
|
25
|
-
|
26
|
-
throw new Error("Missing required authentication information");
|
27
|
-
}
|
28
|
-
const config = getConfluenceRequestConfig(baseUrl, username, authToken);
|
28
|
+
const { pageId, username, content, title } = params;
|
29
|
+
const { baseUrl, authToken } = authParams;
|
30
|
+
const api = getConfluenceApi(baseUrl, username, authToken);
|
29
31
|
// Get current version number
|
30
|
-
const response = yield
|
32
|
+
const response = yield api.get(`/api/v2/pages/${pageId}`);
|
31
33
|
const currVersion = response.data.version.number;
|
32
|
-
|
34
|
+
yield api.put(`/api/v2/pages/${pageId}`, {
|
33
35
|
id: pageId,
|
34
36
|
status: "current",
|
35
37
|
title,
|
@@ -40,7 +42,6 @@ const confluenceUpdatePage = (_a) => __awaiter(void 0, [_a], void 0, function* (
|
|
40
42
|
version: {
|
41
43
|
number: currVersion + 1,
|
42
44
|
},
|
43
|
-
};
|
44
|
-
yield axiosClient_1.axiosClient.put(`/api/v2/pages/${pageId}`, payload, config);
|
45
|
+
});
|
45
46
|
});
|
46
47
|
exports.default = confluenceUpdatePage;
|
@@ -12,27 +12,27 @@ var __importDefault = (this && this.__importDefault) || function (mod) {
|
|
12
12
|
return (mod && mod.__esModule) ? mod : { "default": mod };
|
13
13
|
};
|
14
14
|
Object.defineProperty(exports, "__esModule", { value: true });
|
15
|
-
const
|
16
|
-
const
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
15
|
+
const firecrawl_js_1 = __importDefault(require("@mendable/firecrawl-js"));
|
16
|
+
const types_1 = require("../../autogen/types");
|
17
|
+
const deepResearch = (_a) => __awaiter(void 0, [_a], void 0, function* ({ params, authParams, }) {
|
18
|
+
const { query, maxDepth, maxUrls, timeLimit } = params;
|
19
|
+
const firecrawl = new firecrawl_js_1.default({
|
20
|
+
apiKey: authParams.apiKey,
|
21
|
+
});
|
22
|
+
const result = yield firecrawl.deepResearch(query, {
|
23
|
+
maxDepth,
|
24
|
+
maxUrls,
|
25
|
+
timeLimit,
|
26
|
+
});
|
27
|
+
if (result.success && result.data) {
|
28
|
+
return types_1.firecrawlDeepResearchOutputSchema.parse({
|
29
|
+
finalAnalysis: result.data.finalAnalysis,
|
30
|
+
sources: result.data.sources,
|
24
31
|
});
|
25
|
-
return {
|
26
|
-
statusCode: response.status,
|
27
|
-
headers: response.headers,
|
28
|
-
data: response.data,
|
29
|
-
};
|
30
|
-
}
|
31
|
-
catch (error) {
|
32
|
-
if (axios_1.default.isAxiosError(error)) {
|
33
|
-
throw Error("Axios Error: " + (error.message || "Failed to make API call"));
|
34
|
-
}
|
35
|
-
throw Error("Error: " + (error || "Failed to make API call"));
|
36
32
|
}
|
33
|
+
return {
|
34
|
+
finalAnalysis: "Error",
|
35
|
+
sources: [],
|
36
|
+
};
|
37
37
|
});
|
38
|
-
exports.default =
|
38
|
+
exports.default = deepResearch;
|
@@ -36,7 +36,11 @@ const getSalesforceRecordsByQuery = (_a) => __awaiter(void 0, [_a], void 0, func
|
|
36
36
|
console.error("Error retrieving Salesforce record:", error);
|
37
37
|
return {
|
38
38
|
success: false,
|
39
|
-
error: error instanceof
|
39
|
+
error: error instanceof axiosClient_1.ApiError
|
40
|
+
? error.data.length > 0
|
41
|
+
? error.data[0].message
|
42
|
+
: error.message
|
43
|
+
: "An unknown error occurred",
|
40
44
|
};
|
41
45
|
}
|
42
46
|
});
|
@@ -0,0 +1,23 @@
|
|
1
|
+
declare const axios: any;
|
2
|
+
declare const WORKDAY_BASE_URL = "https://your-workday-url/ccx/service/YOUR_TENANT/Absence_Management/v43.2";
|
3
|
+
declare const TOKEN_URL = "https://your-workday-url/oauth2/YOUR_TENANT/token";
|
4
|
+
declare const CLIENT_ID = "your-client-id";
|
5
|
+
declare const CLIENT_SECRET = "your-client-secret";
|
6
|
+
/**
|
7
|
+
* Fetches an OAuth 2.0 access token from Workday.
|
8
|
+
*/
|
9
|
+
declare function getAccessToken(): Promise<any>;
|
10
|
+
/**
|
11
|
+
* Submits a time-off request to Workday.
|
12
|
+
* @param {Object} params - Time-off details.
|
13
|
+
* @param {string} params.workerId - Worker's ID in Workday.
|
14
|
+
* @param {string} params.startDate - Start date (YYYY-MM-DD).
|
15
|
+
* @param {string} params.endDate - End date (YYYY-MM-DD).
|
16
|
+
* @param {string} params.timeOffType - Time-off type (e.g., "SICK_LEAVE").
|
17
|
+
*/
|
18
|
+
declare function submitTimeOff({ workerId, startDate, endDate, timeOffType }: {
|
19
|
+
workerId: any;
|
20
|
+
startDate: any;
|
21
|
+
endDate: any;
|
22
|
+
timeOffType: any;
|
23
|
+
}): Promise<any>;
|
@@ -0,0 +1,88 @@
|
|
1
|
+
"use strict";
|
2
|
+
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
3
|
+
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
4
|
+
return new (P || (P = Promise))(function (resolve, reject) {
|
5
|
+
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
6
|
+
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
7
|
+
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
8
|
+
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
|
+
});
|
10
|
+
};
|
11
|
+
const axios = require("axios");
|
12
|
+
const WORKDAY_BASE_URL = "https://your-workday-url/ccx/service/YOUR_TENANT/Absence_Management/v43.2";
|
13
|
+
const TOKEN_URL = "https://your-workday-url/oauth2/YOUR_TENANT/token"; // OAuth token endpoint
|
14
|
+
const CLIENT_ID = "your-client-id";
|
15
|
+
const CLIENT_SECRET = "your-client-secret";
|
16
|
+
/**
|
17
|
+
* Fetches an OAuth 2.0 access token from Workday.
|
18
|
+
*/
|
19
|
+
function getAccessToken() {
|
20
|
+
return __awaiter(this, void 0, void 0, function* () {
|
21
|
+
var _a;
|
22
|
+
try {
|
23
|
+
const response = yield axios.post(TOKEN_URL, new URLSearchParams({ grant_type: "client_credentials" }), {
|
24
|
+
auth: {
|
25
|
+
username: CLIENT_ID,
|
26
|
+
password: CLIENT_SECRET
|
27
|
+
},
|
28
|
+
headers: { "Content-Type": "application/x-www-form-urlencoded" }
|
29
|
+
});
|
30
|
+
return response.data.access_token;
|
31
|
+
}
|
32
|
+
catch (error) {
|
33
|
+
console.error("Error fetching access token:", ((_a = error.response) === null || _a === void 0 ? void 0 : _a.data) || error.message);
|
34
|
+
throw error;
|
35
|
+
}
|
36
|
+
});
|
37
|
+
}
|
38
|
+
/**
|
39
|
+
* Submits a time-off request to Workday.
|
40
|
+
* @param {Object} params - Time-off details.
|
41
|
+
* @param {string} params.workerId - Worker's ID in Workday.
|
42
|
+
* @param {string} params.startDate - Start date (YYYY-MM-DD).
|
43
|
+
* @param {string} params.endDate - End date (YYYY-MM-DD).
|
44
|
+
* @param {string} params.timeOffType - Time-off type (e.g., "SICK_LEAVE").
|
45
|
+
*/
|
46
|
+
function submitTimeOff(_a) {
|
47
|
+
return __awaiter(this, arguments, void 0, function* ({ workerId, startDate, endDate, timeOffType }) {
|
48
|
+
var _b;
|
49
|
+
try {
|
50
|
+
const token = yield getAccessToken(); // Get OAuth token
|
51
|
+
const requestBody = {
|
52
|
+
"wd:Enter_Time_Off_Request": {
|
53
|
+
"wd:Worker_Reference": {
|
54
|
+
"wd:ID": [{ "_": workerId, "$": { "wd:type": "WID" } }]
|
55
|
+
},
|
56
|
+
"wd:Time_Off_Entries": [
|
57
|
+
{
|
58
|
+
"wd:Start_Date": startDate,
|
59
|
+
"wd:End_Date": endDate,
|
60
|
+
"wd:Time_Off_Type_Reference": {
|
61
|
+
"wd:ID": [{ "_": timeOffType, "$": { "wd:type": "Time_Off_Type_ID" } }]
|
62
|
+
}
|
63
|
+
}
|
64
|
+
]
|
65
|
+
}
|
66
|
+
};
|
67
|
+
const response = yield axios.post(`${WORKDAY_BASE_URL}/Enter_Time_Off`, requestBody, {
|
68
|
+
headers: {
|
69
|
+
"Authorization": `Bearer ${token}`,
|
70
|
+
"Content-Type": "application/json"
|
71
|
+
}
|
72
|
+
});
|
73
|
+
console.log("Time-off request submitted successfully:", response.data);
|
74
|
+
return response.data;
|
75
|
+
}
|
76
|
+
catch (error) {
|
77
|
+
console.error("Error submitting time-off request:", ((_b = error.response) === null || _b === void 0 ? void 0 : _b.data) || error.message);
|
78
|
+
throw error;
|
79
|
+
}
|
80
|
+
});
|
81
|
+
}
|
82
|
+
// Example Usage:
|
83
|
+
submitTimeOff({
|
84
|
+
workerId: "12345",
|
85
|
+
startDate: "2025-03-10",
|
86
|
+
endDate: "2025-03-12",
|
87
|
+
timeOffType: "SICK_LEAVE"
|
88
|
+
}).then(console.log).catch(console.error);
|
package/package.json
CHANGED
@@ -1,18 +0,0 @@
|
|
1
|
-
"use strict";
|
2
|
-
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
3
|
-
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
4
|
-
return new (P || (P = Promise))(function (resolve, reject) {
|
5
|
-
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
6
|
-
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
7
|
-
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
8
|
-
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
|
-
});
|
10
|
-
};
|
11
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
12
|
-
const fillTemplateAction = (_a) => __awaiter(void 0, [_a], void 0, function* ({ template }) {
|
13
|
-
// Simply return the template without any modification
|
14
|
-
return {
|
15
|
-
result: template,
|
16
|
-
};
|
17
|
-
});
|
18
|
-
exports.default = fillTemplateAction;
|
@@ -1,96 +0,0 @@
|
|
1
|
-
"use strict";
|
2
|
-
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
3
|
-
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
4
|
-
return new (P || (P = Promise))(function (resolve, reject) {
|
5
|
-
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
6
|
-
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
7
|
-
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
8
|
-
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
|
-
});
|
10
|
-
};
|
11
|
-
var __importDefault = (this && this.__importDefault) || function (mod) {
|
12
|
-
return (mod && mod.__esModule) ? mod : { "default": mod };
|
13
|
-
};
|
14
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
15
|
-
const axios_1 = __importDefault(require("axios"));
|
16
|
-
const types_1 = require("../../autogen/types");
|
17
|
-
const INCLUDED_TYPES = [
|
18
|
-
"monument",
|
19
|
-
"museum",
|
20
|
-
"art_gallery",
|
21
|
-
"sculpture",
|
22
|
-
"cultural_landmark",
|
23
|
-
"historical_place",
|
24
|
-
"performing_arts_theater",
|
25
|
-
"university",
|
26
|
-
"aquarium",
|
27
|
-
"botanical_garden",
|
28
|
-
"comedy_club",
|
29
|
-
"park",
|
30
|
-
"movie_theater",
|
31
|
-
"national_park",
|
32
|
-
"garden",
|
33
|
-
"night_club",
|
34
|
-
"tourist_attraction",
|
35
|
-
"water_park",
|
36
|
-
"zoo",
|
37
|
-
"bar",
|
38
|
-
"restaurant",
|
39
|
-
"food_court",
|
40
|
-
"bakery",
|
41
|
-
"cafe",
|
42
|
-
"coffee_shop",
|
43
|
-
"pub",
|
44
|
-
"wine_bar",
|
45
|
-
"spa",
|
46
|
-
"beach",
|
47
|
-
"market",
|
48
|
-
"shopping_mall",
|
49
|
-
"stadium",
|
50
|
-
];
|
51
|
-
const nearbysearch = (_a) => __awaiter(void 0, [_a], void 0, function* ({ params, authParams, }) {
|
52
|
-
const url = `https://places.googleapis.com/v1/places:searchNearby`;
|
53
|
-
const fieldMask = [
|
54
|
-
"places.displayName",
|
55
|
-
"places.formattedAddress",
|
56
|
-
"places.priceLevel",
|
57
|
-
"places.rating",
|
58
|
-
"places.primaryTypeDisplayName",
|
59
|
-
"places.editorialSummary",
|
60
|
-
"places.regularOpeningHours",
|
61
|
-
].join(",");
|
62
|
-
const response = yield axios_1.default.post(url, {
|
63
|
-
maxResultCount: 20,
|
64
|
-
includedTypes: INCLUDED_TYPES,
|
65
|
-
locationRestriction: {
|
66
|
-
circle: {
|
67
|
-
center: {
|
68
|
-
latitude: params.latitude,
|
69
|
-
longitude: params.longitude,
|
70
|
-
},
|
71
|
-
radius: 10000,
|
72
|
-
},
|
73
|
-
},
|
74
|
-
}, {
|
75
|
-
headers: {
|
76
|
-
"X-Goog-Api-Key": authParams.apiKey,
|
77
|
-
"X-Goog-FieldMask": fieldMask,
|
78
|
-
"Content-Type": "application/json",
|
79
|
-
},
|
80
|
-
});
|
81
|
-
return types_1.googlemapsNearbysearchOutputSchema.parse({
|
82
|
-
results: response.data.places.map((place) => {
|
83
|
-
var _a, _b;
|
84
|
-
return ({
|
85
|
-
name: place.displayName.text,
|
86
|
-
address: place.formattedAddress,
|
87
|
-
priceLevel: place.priceLevel,
|
88
|
-
rating: place.rating,
|
89
|
-
primaryType: place.primaryTypeDisplayName.text,
|
90
|
-
editorialSummary: ((_a = place.editorialSummary) === null || _a === void 0 ? void 0 : _a.text) || "",
|
91
|
-
openingHours: ((_b = place.regularOpeningHours) === null || _b === void 0 ? void 0 : _b.weekdayDescriptions.join("\n")) || "",
|
92
|
-
});
|
93
|
-
}),
|
94
|
-
});
|
95
|
-
});
|
96
|
-
exports.default = nearbysearch;
|
@@ -1,154 +0,0 @@
|
|
1
|
-
"use strict";
|
2
|
-
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
3
|
-
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
4
|
-
return new (P || (P = Promise))(function (resolve, reject) {
|
5
|
-
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
6
|
-
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
7
|
-
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
8
|
-
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
|
-
});
|
10
|
-
};
|
11
|
-
var __importDefault = (this && this.__importDefault) || function (mod) {
|
12
|
-
return (mod && mod.__esModule) ? mod : { "default": mod };
|
13
|
-
};
|
14
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
15
|
-
const snowflake_sdk_1 = __importDefault(require("snowflake-sdk"));
|
16
|
-
const crypto_1 = __importDefault(require("crypto"));
|
17
|
-
const client_s3_1 = require("@aws-sdk/client-s3");
|
18
|
-
const s3_request_presigner_1 = require("@aws-sdk/s3-request-presigner");
|
19
|
-
const uuid_1 = require("uuid");
|
20
|
-
// Only log errors.
|
21
|
-
snowflake_sdk_1.default.configure({ logLevel: "ERROR" });
|
22
|
-
const runSnowflakeQueryWriteResultsToS3 = (_a) => __awaiter(void 0, [_a], void 0, function* ({ params, authParams, }) {
|
23
|
-
const { databaseName, warehouse, query, user, accountName, s3BucketName, s3Region, outputFormat = "json" } = params;
|
24
|
-
const { apiKey: privateKey, awsAccessKeyId, awsSecretAccessKey } = authParams;
|
25
|
-
if (!privateKey) {
|
26
|
-
throw new Error("Snowflake private key is required");
|
27
|
-
}
|
28
|
-
if (!awsAccessKeyId || !awsSecretAccessKey) {
|
29
|
-
throw new Error("AWS credentials are required");
|
30
|
-
}
|
31
|
-
if (!accountName || !user || !databaseName || !warehouse || !query || !s3BucketName) {
|
32
|
-
throw new Error("Missing required parameters for Snowflake query or S3 destination");
|
33
|
-
}
|
34
|
-
const getPrivateKeyCorrectFormat = (privateKey) => {
|
35
|
-
const buffer = Buffer.from(privateKey);
|
36
|
-
const privateKeyObject = crypto_1.default.createPrivateKey({
|
37
|
-
key: buffer,
|
38
|
-
format: "pem",
|
39
|
-
passphrase: "password",
|
40
|
-
});
|
41
|
-
const privateKeyCorrectFormat = privateKeyObject.export({
|
42
|
-
format: "pem",
|
43
|
-
type: "pkcs8",
|
44
|
-
});
|
45
|
-
return privateKeyCorrectFormat.toString();
|
46
|
-
};
|
47
|
-
const executeQueryAndFormatData = () => __awaiter(void 0, void 0, void 0, function* () {
|
48
|
-
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
49
|
-
const queryResults = yield new Promise((resolve, reject) => {
|
50
|
-
connection.execute({
|
51
|
-
sqlText: query,
|
52
|
-
complete: (err, stmt, rows) => {
|
53
|
-
if (err) {
|
54
|
-
return reject(err);
|
55
|
-
}
|
56
|
-
return resolve(rows || []);
|
57
|
-
},
|
58
|
-
});
|
59
|
-
});
|
60
|
-
// Format the results based on the output format
|
61
|
-
let formattedData;
|
62
|
-
if (outputFormat.toLowerCase() === "csv") {
|
63
|
-
if (queryResults.length === 0) {
|
64
|
-
formattedData = "";
|
65
|
-
}
|
66
|
-
else {
|
67
|
-
const headers = Object.keys(queryResults[0]).join(",");
|
68
|
-
const rows = queryResults.map(row => Object.values(row)
|
69
|
-
.map(value => (typeof value === "object" && value !== null ? JSON.stringify(value) : value))
|
70
|
-
.join(","));
|
71
|
-
formattedData = [headers, ...rows].join("\n");
|
72
|
-
}
|
73
|
-
}
|
74
|
-
else {
|
75
|
-
// Default to JSON
|
76
|
-
formattedData = JSON.stringify(queryResults, null, 2);
|
77
|
-
}
|
78
|
-
return { formattedData, resultsLength: queryResults.length };
|
79
|
-
});
|
80
|
-
const uploadToS3AndGetURL = (formattedData) => __awaiter(void 0, void 0, void 0, function* () {
|
81
|
-
// Create S3 client
|
82
|
-
const s3Client = new client_s3_1.S3Client({
|
83
|
-
region: s3Region,
|
84
|
-
credentials: {
|
85
|
-
accessKeyId: awsAccessKeyId,
|
86
|
-
secretAccessKey: awsSecretAccessKey,
|
87
|
-
},
|
88
|
-
});
|
89
|
-
const contentType = outputFormat.toLowerCase() === "csv" ? "text/csv" : "application/json";
|
90
|
-
const fileExtension = outputFormat.toLowerCase() === "csv" ? "csv" : "json";
|
91
|
-
const finalKey = `${databaseName}/${(0, uuid_1.v4)()}.${fileExtension}`;
|
92
|
-
// Upload to S3 without ACL
|
93
|
-
const uploadCommand = new client_s3_1.PutObjectCommand({
|
94
|
-
Bucket: s3BucketName,
|
95
|
-
Key: finalKey,
|
96
|
-
Body: formattedData,
|
97
|
-
ContentType: contentType,
|
98
|
-
});
|
99
|
-
yield s3Client.send(uploadCommand);
|
100
|
-
// Generate a presigned URL (valid for an hour)
|
101
|
-
const getObjectCommand = new client_s3_1.GetObjectCommand({
|
102
|
-
Bucket: s3BucketName,
|
103
|
-
Key: finalKey,
|
104
|
-
});
|
105
|
-
const presignedUrl = yield (0, s3_request_presigner_1.getSignedUrl)(s3Client, getObjectCommand, { expiresIn: 3600 });
|
106
|
-
return presignedUrl;
|
107
|
-
});
|
108
|
-
// Process the private key
|
109
|
-
const privateKeyCorrectFormatString = getPrivateKeyCorrectFormat(privateKey);
|
110
|
-
// Set up a connection using snowflake-sdk
|
111
|
-
const connection = snowflake_sdk_1.default.createConnection({
|
112
|
-
account: accountName,
|
113
|
-
username: user,
|
114
|
-
privateKey: privateKeyCorrectFormatString,
|
115
|
-
authenticator: "SNOWFLAKE_JWT",
|
116
|
-
role: "ACCOUNTADMIN",
|
117
|
-
warehouse: warehouse,
|
118
|
-
database: databaseName,
|
119
|
-
});
|
120
|
-
try {
|
121
|
-
// Connect to Snowflake
|
122
|
-
yield new Promise((resolve, reject) => {
|
123
|
-
connection.connect((err, conn) => {
|
124
|
-
if (err) {
|
125
|
-
console.error("Unable to connect to Snowflake:", err.message);
|
126
|
-
return reject(err);
|
127
|
-
}
|
128
|
-
resolve(conn);
|
129
|
-
});
|
130
|
-
});
|
131
|
-
const { formattedData, resultsLength } = yield executeQueryAndFormatData();
|
132
|
-
const presignedUrl = yield uploadToS3AndGetURL(formattedData);
|
133
|
-
// Return fields to match schema definition
|
134
|
-
connection.destroy(err => {
|
135
|
-
if (err) {
|
136
|
-
console.log("Failed to disconnect from Snowflake:", err);
|
137
|
-
}
|
138
|
-
});
|
139
|
-
return {
|
140
|
-
bucketUrl: presignedUrl,
|
141
|
-
message: `Query results successfully written to S3. URL valid for 1 hour.`,
|
142
|
-
rowCount: resultsLength,
|
143
|
-
};
|
144
|
-
}
|
145
|
-
catch (error) {
|
146
|
-
connection.destroy(err => {
|
147
|
-
if (err) {
|
148
|
-
console.log("Failed to disconnect from Snowflake:", err);
|
149
|
-
}
|
150
|
-
});
|
151
|
-
throw Error(`An error occurred: ${error}`);
|
152
|
-
}
|
153
|
-
});
|
154
|
-
exports.default = runSnowflakeQueryWriteResultsToS3;
|
@@ -1,45 +0,0 @@
|
|
1
|
-
"use strict";
|
2
|
-
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
3
|
-
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
4
|
-
return new (P || (P = Promise))(function (resolve, reject) {
|
5
|
-
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
6
|
-
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
7
|
-
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
8
|
-
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
9
|
-
});
|
10
|
-
};
|
11
|
-
var __importDefault = (this && this.__importDefault) || function (mod) {
|
12
|
-
return (mod && mod.__esModule) ? mod : { "default": mod };
|
13
|
-
};
|
14
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
15
|
-
const firecrawl_js_1 = __importDefault(require("@mendable/firecrawl-js"));
|
16
|
-
const scrapeTweetDataWithNitter = (_a) => __awaiter(void 0, [_a], void 0, function* ({ params, authParams, }) {
|
17
|
-
const tweetUrlRegex = /^(?:https?:\/\/)?(?:www\.)?(?:twitter\.com|x\.com)\/([a-zA-Z0-9_]+)\/status\/(\d+)(?:\?.*)?$/;
|
18
|
-
if (!tweetUrlRegex.test(params.tweetUrl)) {
|
19
|
-
throw new Error("Invalid tweet URL. Expected format: https://twitter.com/username/status/id or https://x.com/username/status/id");
|
20
|
-
}
|
21
|
-
const nitterUrl = params.tweetUrl.replace(/^(?:https?:\/\/)?(?:www\.)?(?:twitter\.com|x\.com)/i, "https://nitter.net");
|
22
|
-
// Initialize Firecrawl
|
23
|
-
if (!authParams.apiKey) {
|
24
|
-
throw new Error("API key is required for X+Nitter+Firecrawl");
|
25
|
-
}
|
26
|
-
const firecrawl = new firecrawl_js_1.default({
|
27
|
-
apiKey: authParams.apiKey,
|
28
|
-
});
|
29
|
-
try {
|
30
|
-
// Scrape the Nitter URL
|
31
|
-
const result = yield firecrawl.scrapeUrl(nitterUrl);
|
32
|
-
if (!result.success) {
|
33
|
-
throw new Error(`Failed to scrape tweet: ${result.error || "Unknown error"}`);
|
34
|
-
}
|
35
|
-
// Extract the tweet text from the scraped content - simple approach - in practice, you might need more robust parsing based on nitter html structure
|
36
|
-
const tweetContent = result.markdown;
|
37
|
-
return {
|
38
|
-
text: tweetContent || "Error scraping with firecrawl",
|
39
|
-
};
|
40
|
-
}
|
41
|
-
catch (error) {
|
42
|
-
throw new Error(`Error scraping tweet: ${error instanceof Error ? error.message : error}`);
|
43
|
-
}
|
44
|
-
});
|
45
|
-
exports.default = scrapeTweetDataWithNitter;
|