n8n-nodes-github-copilot 3.2.6 → 3.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -43,7 +43,7 @@ class GitHubCopilotChatAPI {
43
43
  const advancedOptions = this.getNodeParameter('advancedOptions', i, {});
44
44
  const includeMedia = this.getNodeParameter('includeMedia', i, false);
45
45
  if (includeMedia) {
46
- const validation = (0, modelCapabilities_1.validateModelCapabilities)(model, true, true);
46
+ const validation = (0, modelCapabilities_1.validateModelCapabilities)(model, true, false);
47
47
  if (!validation.isValid) {
48
48
  throw new Error(validation.errorMessage || 'Model validation failed');
49
49
  }
@@ -83,16 +83,10 @@ class GitHubCopilotChatAPI {
83
83
  },
84
84
  });
85
85
  }
86
- else if (mediaResult.type === 'audio' && mediaResult.dataUrl) {
87
- contentArray.push({
88
- type: 'text',
89
- text: `Please analyze this audio file: ${mediaResult.description}\n\nAudio data (base64): ${mediaResult.dataUrl}`,
90
- });
91
- }
92
86
  else {
93
87
  contentArray.push({
94
88
  type: 'text',
95
- text: `[Media processing failed: ${mediaResult.description}]`,
89
+ text: `[Image processing failed: ${mediaResult.description}]`,
96
90
  });
97
91
  }
98
92
  }
@@ -24,12 +24,12 @@ exports.nodeProperties = [
24
24
  {
25
25
  name: 'GPT-5',
26
26
  value: 'gpt-5',
27
- description: 'OpenAI GPT-5 (Latest and most capable) ✓ Images ✓ Audio',
27
+ description: 'OpenAI GPT-5 (Latest and most capable) ✓ Images',
28
28
  },
29
29
  {
30
30
  name: 'GPT-5 Mini',
31
31
  value: 'gpt-5-mini',
32
- description: 'OpenAI GPT-5 Mini (Faster, cost-effective) ✓ Images ✓ Audio',
32
+ description: 'OpenAI GPT-5 Mini (Faster, cost-effective) ✓ Images',
33
33
  },
34
34
  {
35
35
  name: 'Claude Opus 4.1',
@@ -39,7 +39,7 @@ exports.nodeProperties = [
39
39
  {
40
40
  name: 'Gemini 2.5 Pro',
41
41
  value: 'gemini-2.5-pro',
42
- description: 'Google Gemini 2.5 Pro (Multimodal capabilities) ✓ Images ✓ Audio',
42
+ description: 'Google Gemini 2.5 Pro (Multimodal capabilities) ✓ Images',
43
43
  },
44
44
  {
45
45
  name: 'Grok Code Fast 1',
@@ -49,7 +49,7 @@ exports.nodeProperties = [
49
49
  {
50
50
  name: 'GPT-4.1 Copilot',
51
51
  value: 'gpt-4.1-copilot',
52
- description: 'OpenAI GPT-4.1 specialized for coding assistance ✓ Images ✓ Audio',
52
+ description: 'OpenAI GPT-4.1 specialized for coding assistance ✓ Images',
53
53
  },
54
54
  ],
55
55
  default: 'gpt-5-mini',
@@ -78,14 +78,14 @@ exports.nodeProperties = [
78
78
  description: 'System message to set the behavior of the AI model',
79
79
  },
80
80
  {
81
- displayName: 'Include Media',
81
+ displayName: 'Include Image',
82
82
  name: 'includeMedia',
83
83
  type: 'boolean',
84
84
  default: false,
85
- description: 'Whether to include a media file in the message. Supported: Images (PNG, JPEG, GIF, WebP) and Audio files. Type is auto-detected.',
85
+ description: 'Whether to include an image file in the message. Supported formats: PNG, JPEG, GIF, WebP.',
86
86
  },
87
87
  {
88
- displayName: 'Media Source',
88
+ displayName: 'Image Source',
89
89
  name: 'mediaSource',
90
90
  type: 'options',
91
91
  displayOptions: {
@@ -97,12 +97,12 @@ exports.nodeProperties = [
97
97
  {
98
98
  name: 'Manual Input',
99
99
  value: 'manual',
100
- description: 'Provide media as base64 string or file path',
100
+ description: 'Provide image as base64 string or file path',
101
101
  },
102
102
  {
103
103
  name: 'URL',
104
104
  value: 'url',
105
- description: 'Download media from URL',
105
+ description: 'Download image from URL',
106
106
  },
107
107
  {
108
108
  name: 'Binary Data',
@@ -111,10 +111,10 @@ exports.nodeProperties = [
111
111
  },
112
112
  ],
113
113
  default: 'manual',
114
- description: 'Source of the media data',
114
+ description: 'Source of the image data',
115
115
  },
116
116
  {
117
- displayName: 'Media File',
117
+ displayName: 'Image File',
118
118
  name: 'mediaFile',
119
119
  type: 'string',
120
120
  displayOptions: {
@@ -125,10 +125,10 @@ exports.nodeProperties = [
125
125
  },
126
126
  default: '',
127
127
  placeholder: 'Paste base64 string or file path',
128
- description: 'Media file as base64 string or file path (Images: PNG/JPEG/GIF/WebP, Audio: MP3/WAV/M4A)',
128
+ description: 'Image file as base64 string or file path. Supported formats: PNG, JPEG, GIF, WebP.',
129
129
  },
130
130
  {
131
- displayName: 'Media URL',
131
+ displayName: 'Image URL',
132
132
  name: 'mediaUrl',
133
133
  type: 'string',
134
134
  displayOptions: {
@@ -138,11 +138,11 @@ exports.nodeProperties = [
138
138
  },
139
139
  },
140
140
  default: '',
141
- placeholder: 'https://example.com/file.jpg',
142
- description: 'URL of the media file to download and include (Images: PNG/JPEG/GIF/WebP, Audio: MP3/WAV/M4A)',
141
+ placeholder: 'https://example.com/image.jpg',
142
+ description: 'URL of the image file to download. Supported formats: PNG, JPEG, GIF, WebP.',
143
143
  },
144
144
  {
145
- displayName: 'Media Binary Property',
145
+ displayName: 'Image Binary Property',
146
146
  name: 'mediaBinaryProperty',
147
147
  type: 'string',
148
148
  displayOptions: {
@@ -153,7 +153,7 @@ exports.nodeProperties = [
153
153
  },
154
154
  default: 'data',
155
155
  placeholder: 'data',
156
- description: 'Name of the binary property containing the media file',
156
+ description: 'Name of the binary property containing the image file',
157
157
  },
158
158
  {
159
159
  displayName: 'Advanced Options',
@@ -1,4 +1,3 @@
1
1
  export * from './types';
2
2
  export * from './helpers';
3
- export * from './audioProcessor';
4
3
  export * from './imageProcessor';
@@ -16,5 +16,4 @@ var __exportStar = (this && this.__exportStar) || function(m, exports) {
16
16
  Object.defineProperty(exports, "__esModule", { value: true });
17
17
  __exportStar(require("./types"), exports);
18
18
  __exportStar(require("./helpers"), exports);
19
- __exportStar(require("./audioProcessor"), exports);
20
19
  __exportStar(require("./imageProcessor"), exports);
@@ -1,12 +1,11 @@
1
1
  import { IExecuteFunctions } from 'n8n-workflow';
2
2
  export declare function processMediaFile(context: IExecuteFunctions, itemIndex: number, source: 'manual' | 'url' | 'binary', mediaFile?: string, mediaUrl?: string, binaryProperty?: string): Promise<{
3
- type: 'image' | 'audio' | 'unknown';
3
+ type: 'image' | 'unknown';
4
4
  dataUrl?: string;
5
5
  description: string;
6
6
  mimeType: string;
7
7
  }>;
8
8
  export declare function isImageMimeType(mimeType: string): boolean;
9
- export declare function isAudioMimeType(mimeType: string): boolean;
10
9
  export declare function validateImageFormat(mimeType: string): {
11
10
  isValid: boolean;
12
11
  error?: string;
@@ -1,151 +1,25 @@
1
1
  "use strict";
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
- exports.suggestImageConversion = exports.getFileExtensionFromMimeType = exports.validateImageFormat = exports.isAudioMimeType = exports.isImageMimeType = exports.processMediaFile = void 0;
3
+ exports.suggestImageConversion = exports.getFileExtensionFromMimeType = exports.validateImageFormat = exports.isImageMimeType = exports.processMediaFile = void 0;
4
4
  const index_1 = require("./index");
5
- async function detectMimeType(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty) {
6
- var _a, _b;
7
- try {
8
- if (source === 'binary') {
9
- const binaryData = context.getInputData()[itemIndex].binary;
10
- if (binaryData && binaryProperty && binaryData[binaryProperty]) {
11
- return binaryData[binaryProperty].mimeType || 'application/octet-stream';
12
- }
13
- }
14
- else if (source === 'url' && mediaUrl) {
15
- const extension = (_a = mediaUrl.split('.').pop()) === null || _a === void 0 ? void 0 : _a.toLowerCase();
16
- const extToMime = {
17
- 'png': 'image/png',
18
- 'jpg': 'image/jpeg',
19
- 'jpeg': 'image/jpeg',
20
- 'gif': 'image/gif',
21
- 'webp': 'image/webp',
22
- 'mp3': 'audio/mpeg',
23
- 'wav': 'audio/wav',
24
- 'm4a': 'audio/mp4',
25
- 'ogg': 'audio/ogg',
26
- 'mp4': 'video/mp4'
27
- };
28
- return extToMime[extension || ''] || 'application/octet-stream';
29
- }
30
- else if (source === 'manual' && mediaFile) {
31
- if (mediaFile.startsWith('data:')) {
32
- const mimeMatch = mediaFile.match(/data:([^;]+)/);
33
- if (mimeMatch)
34
- return mimeMatch[1];
35
- }
36
- const extension = (_b = mediaFile.split('.').pop()) === null || _b === void 0 ? void 0 : _b.toLowerCase();
37
- const extToMime = {
38
- 'png': 'image/png',
39
- 'jpg': 'image/jpeg',
40
- 'jpeg': 'image/jpeg',
41
- 'gif': 'image/gif',
42
- 'webp': 'image/webp',
43
- 'mp3': 'audio/mpeg',
44
- 'wav': 'audio/wav',
45
- 'm4a': 'audio/mp4',
46
- 'ogg': 'audio/ogg',
47
- 'mp4': 'video/mp4'
48
- };
49
- return extToMime[extension || ''] || 'application/octet-stream';
50
- }
51
- }
52
- catch (error) {
53
- }
54
- return 'application/octet-stream';
55
- }
56
5
  async function processMediaFile(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty) {
57
6
  try {
58
- const detectedMimeType = await detectMimeType(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
59
- const isLikelyImage = isImageMimeType(detectedMimeType);
60
- const isLikelyAudio = isAudioMimeType(detectedMimeType);
61
- if (isLikelyAudio && !isLikelyImage) {
62
- try {
63
- const audioResult = await (0, index_1.processAudioFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
64
- return {
65
- type: 'audio',
66
- dataUrl: `data:${audioResult.mimeType};base64,${audioResult.data}`,
67
- description: `Audio file: ${audioResult.filename} (${Math.round(audioResult.size / 1024)}KB, ${audioResult.mimeType})`,
68
- mimeType: audioResult.mimeType,
69
- };
70
- }
71
- catch (audioError) {
72
- throw new Error(`Audio processing failed: ${audioError instanceof Error ? audioError.message : 'Unknown error'}`);
73
- }
74
- }
75
- else if (isLikelyImage && !isLikelyAudio) {
76
- try {
77
- const imageResult = await (0, index_1.processImageFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
78
- const formatValidation = validateImageFormat(imageResult.mimeType);
79
- if (!formatValidation.isValid) {
80
- throw new Error(suggestImageConversion(imageResult.mimeType));
81
- }
82
- return {
83
- type: 'image',
84
- dataUrl: `data:${imageResult.mimeType};base64,${imageResult.data}`,
85
- description: `Image file: ${imageResult.filename} (${Math.round(imageResult.size / 1024)}KB, ${imageResult.mimeType})`,
86
- mimeType: imageResult.mimeType,
87
- };
88
- }
89
- catch (imageError) {
90
- throw new Error(`Image processing failed: ${imageError instanceof Error ? imageError.message : 'Unknown error'}`);
91
- }
92
- }
93
- else {
94
- if (detectedMimeType.includes('mpeg') || detectedMimeType.includes('mp3')) {
95
- try {
96
- const audioResult = await (0, index_1.processAudioFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
97
- return {
98
- type: 'audio',
99
- dataUrl: `data:${audioResult.mimeType};base64,${audioResult.data}`,
100
- description: `Audio file: ${audioResult.filename} (${Math.round(audioResult.size / 1024)}KB, ${audioResult.mimeType})`,
101
- mimeType: audioResult.mimeType,
102
- };
103
- }
104
- catch (audioError) {
105
- throw new Error(`MPEG audio processing failed: ${audioError instanceof Error ? audioError.message : 'Unknown error'}`);
106
- }
107
- }
108
- else {
109
- try {
110
- const imageResult = await (0, index_1.processImageFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
111
- const formatValidation = validateImageFormat(imageResult.mimeType);
112
- if (!formatValidation.isValid) {
113
- throw new Error(suggestImageConversion(imageResult.mimeType));
114
- }
115
- return {
116
- type: 'image',
117
- dataUrl: `data:${imageResult.mimeType};base64,${imageResult.data}`,
118
- description: `Image file: ${imageResult.filename} (${Math.round(imageResult.size / 1024)}KB, ${imageResult.mimeType})`,
119
- mimeType: imageResult.mimeType,
120
- };
121
- }
122
- catch (imageError) {
123
- try {
124
- const audioResult = await (0, index_1.processAudioFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
125
- return {
126
- type: 'audio',
127
- dataUrl: `data:${audioResult.mimeType};base64,${audioResult.data}`,
128
- description: `Audio file: ${audioResult.filename} (${Math.round(audioResult.size / 1024)}KB, ${audioResult.mimeType})`,
129
- mimeType: audioResult.mimeType,
130
- };
131
- }
132
- catch (audioError) {
133
- const supportedImageFormats = ['PNG', 'JPEG', 'GIF', 'WebP'];
134
- const supportedAudioFormats = ['MP3', 'WAV', 'M4A', 'OGG'];
135
- throw new Error(`File is neither a valid image nor audio file (detected MIME: ${detectedMimeType}).\n` +
136
- `Supported image formats: ${supportedImageFormats.join(', ')}\n` +
137
- `Supported audio formats: ${supportedAudioFormats.join(', ')}\n` +
138
- `Image error: ${imageError instanceof Error ? imageError.message : 'Unknown'}\n` +
139
- `Audio error: ${audioError instanceof Error ? audioError.message : 'Unknown'}`);
140
- }
141
- }
142
- }
7
+ const imageResult = await (0, index_1.processImageFile)(context, itemIndex, source, mediaFile, mediaUrl, binaryProperty);
8
+ const formatValidation = validateImageFormat(imageResult.mimeType);
9
+ if (!formatValidation.isValid) {
10
+ throw new Error(suggestImageConversion(imageResult.mimeType));
143
11
  }
12
+ return {
13
+ type: 'image',
14
+ dataUrl: `data:${imageResult.mimeType};base64,${imageResult.data}`,
15
+ description: `Image file: ${imageResult.filename} (${Math.round(imageResult.size / 1024)}KB, ${imageResult.mimeType})`,
16
+ mimeType: imageResult.mimeType,
17
+ };
144
18
  }
145
19
  catch (error) {
146
20
  return {
147
21
  type: 'unknown',
148
- description: `Error processing media file: ${error instanceof Error ? error.message : 'Unknown error'}`,
22
+ description: `Error processing image file: ${error instanceof Error ? error.message : 'Unknown error'}`,
149
23
  mimeType: 'unknown',
150
24
  };
151
25
  }
@@ -162,12 +36,6 @@ function isImageMimeType(mimeType) {
162
36
  return supportedFormats.includes(mimeType.toLowerCase());
163
37
  }
164
38
  exports.isImageMimeType = isImageMimeType;
165
- function isAudioMimeType(mimeType) {
166
- return mimeType.startsWith('audio/') ||
167
- mimeType === 'application/ogg' ||
168
- mimeType === 'video/mp4';
169
- }
170
- exports.isAudioMimeType = isAudioMimeType;
171
39
  function validateImageFormat(mimeType) {
172
40
  if (!isImageMimeType(mimeType)) {
173
41
  const supportedFormats = ['PNG', 'JPEG', 'GIF', 'WebP'];
@@ -4,21 +4,21 @@ exports.getModelInfo = exports.getSupportedModels = exports.validateModelCapabil
4
4
  exports.MODEL_CAPABILITIES = {
5
5
  'gpt-5': {
6
6
  supportsImages: true,
7
- supportsAudio: true,
7
+ supportsAudio: false,
8
8
  maxContextTokens: 200000,
9
- description: 'OpenAI GPT-5 with image support and audio analysis via GitHub Copilot API',
9
+ description: 'OpenAI GPT-5 with image support via GitHub Copilot API',
10
10
  },
11
11
  'gpt-5-mini': {
12
12
  supportsImages: true,
13
- supportsAudio: true,
13
+ supportsAudio: false,
14
14
  maxContextTokens: 128000,
15
- description: 'OpenAI GPT-5 Mini with image support and audio analysis via GitHub Copilot API',
15
+ description: 'OpenAI GPT-5 Mini with image support via GitHub Copilot API',
16
16
  },
17
17
  'gpt-4.1-copilot': {
18
18
  supportsImages: true,
19
- supportsAudio: true,
19
+ supportsAudio: false,
20
20
  maxContextTokens: 128000,
21
- description: 'OpenAI GPT-4.1 with image support and audio analysis via GitHub Copilot API',
21
+ description: 'OpenAI GPT-4.1 with image support via GitHub Copilot API',
22
22
  },
23
23
  'claude-opus-4.1': {
24
24
  supportsImages: false,
@@ -34,9 +34,9 @@ exports.MODEL_CAPABILITIES = {
34
34
  },
35
35
  'gemini-2.5-pro': {
36
36
  supportsImages: true,
37
- supportsAudio: true,
37
+ supportsAudio: false,
38
38
  maxContextTokens: 1000000,
39
- description: 'Google Gemini 2.5 Pro with multimodal support via GitHub Copilot API',
39
+ description: 'Google Gemini 2.5 Pro with image support via GitHub Copilot API',
40
40
  },
41
41
  'gemini-2.0-flash': {
42
42
  supportsImages: true,
@@ -0,0 +1,5 @@
1
+ import { IExecuteFunctions, INodeExecutionData, INodeType, INodeTypeDescription } from 'n8n-workflow';
2
+ export declare class N8nAiAgent implements INodeType {
3
+ description: INodeTypeDescription;
4
+ execute(this: IExecuteFunctions): Promise<INodeExecutionData[][]>;
5
+ }
@@ -0,0 +1,214 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.N8nAiAgent = void 0;
4
+ const n8n_workflow_1 = require("n8n-workflow");
5
+ const nodeProperties_1 = require("./nodeProperties");
6
+ const mediaDetection_1 = require("../GitHubCopilotChatAPI/utils/mediaDetection");
7
+ class N8nAiAgent {
8
+ constructor() {
9
+ this.description = {
10
+ displayName: 'N8N AI Agent',
11
+ name: 'n8nAiAgent',
12
+ icon: 'file:n8n-ai.svg',
13
+ group: ['AI'],
14
+ version: 1,
15
+ subtitle: '={{$parameter["operation"] + ": " + $parameter["model"]}}',
16
+ description: 'Connect to N8N AI Agent service for advanced AI capabilities with tool calling and memory',
17
+ defaults: {
18
+ name: 'N8N AI Agent',
19
+ },
20
+ inputs: ["main"],
21
+ outputs: ["main"],
22
+ credentials: [
23
+ {
24
+ name: 'n8nApi',
25
+ required: true,
26
+ },
27
+ ],
28
+ properties: nodeProperties_1.nodeProperties,
29
+ };
30
+ }
31
+ async execute() {
32
+ const items = this.getInputData();
33
+ const returnData = [];
34
+ for (let i = 0; i < items.length; i++) {
35
+ try {
36
+ const operation = this.getNodeParameter('operation', i);
37
+ const model = this.getNodeParameter('model', i);
38
+ if (operation === 'chat') {
39
+ const message = this.getNodeParameter('message', i);
40
+ const includeMedia = this.getNodeParameter('includeMedia', i, false);
41
+ const messages = [];
42
+ const systemMessage = this.getNodeParameter('systemMessage', i, '');
43
+ if (systemMessage) {
44
+ messages.push({
45
+ role: 'system',
46
+ content: systemMessage,
47
+ });
48
+ }
49
+ const messageContent = [
50
+ {
51
+ type: 'text',
52
+ text: message,
53
+ },
54
+ ];
55
+ if (includeMedia) {
56
+ const mediaSource = this.getNodeParameter('mediaSource', i);
57
+ const processedMedia = await (0, mediaDetection_1.processMediaFile)(this, i, mediaSource, this.getNodeParameter('mediaFile', i, ''), this.getNodeParameter('mediaUrl', i, ''), this.getNodeParameter('mediaProperty', i, ''));
58
+ if (processedMedia && processedMedia.dataUrl) {
59
+ messageContent.push({
60
+ type: 'image_url',
61
+ image_url: {
62
+ url: processedMedia.dataUrl,
63
+ },
64
+ });
65
+ }
66
+ }
67
+ messages.push({
68
+ role: 'user',
69
+ content: messageContent,
70
+ });
71
+ const includeHistory = this.getNodeParameter('includeHistory', i, false);
72
+ if (includeHistory) {
73
+ const historyMessages = this.getNodeParameter('conversationHistory', i, []);
74
+ messages.splice(-1, 0, ...historyMessages);
75
+ }
76
+ const advancedOptions = {};
77
+ const maxTokens = this.getNodeParameter('maxTokens', i, 1000);
78
+ const temperature = this.getNodeParameter('temperature', i, 0.7);
79
+ const enableTools = this.getNodeParameter('enableTools', i, false);
80
+ if (maxTokens > 0) {
81
+ advancedOptions.max_tokens = maxTokens;
82
+ }
83
+ advancedOptions.temperature = temperature;
84
+ if (enableTools) {
85
+ const toolsConfig = this.getNodeParameter('toolsConfig', i, {});
86
+ if (toolsConfig.tools && Array.isArray(toolsConfig.tools)) {
87
+ advancedOptions.tools = toolsConfig.tools;
88
+ }
89
+ }
90
+ const requestBody = {
91
+ model,
92
+ messages,
93
+ stream: false,
94
+ ...advancedOptions,
95
+ };
96
+ const response = await makeN8nAiAgentRequest(this, '/chat', requestBody, includeMedia);
97
+ const result = {
98
+ message: response.response || response.message || '',
99
+ model,
100
+ operation,
101
+ usage: response.usage || null,
102
+ tool_calls: response.tool_calls || null,
103
+ memory: response.memory || null,
104
+ finish_reason: response.finish_reason || 'completed',
105
+ };
106
+ returnData.push({
107
+ json: result,
108
+ pairedItem: { item: i },
109
+ });
110
+ }
111
+ else if (operation === 'tools') {
112
+ const toolName = this.getNodeParameter('toolName', i);
113
+ const toolArguments = this.getNodeParameter('toolArguments', i, {});
114
+ const context = this.getNodeParameter('context', i, '');
115
+ const requestBody = {
116
+ tool: toolName,
117
+ arguments: toolArguments,
118
+ context,
119
+ model,
120
+ };
121
+ const response = await makeN8nAiAgentRequest(this, '/tools', requestBody, false);
122
+ const result = {
123
+ tool_name: toolName,
124
+ result: response.result || response.response,
125
+ execution_time: response.execution_time || null,
126
+ operation,
127
+ model,
128
+ };
129
+ returnData.push({
130
+ json: result,
131
+ pairedItem: { item: i },
132
+ });
133
+ }
134
+ else if (operation === 'memory') {
135
+ const memoryAction = this.getNodeParameter('memoryAction', i);
136
+ const sessionId = this.getNodeParameter('sessionId', i, '');
137
+ const requestBody = {
138
+ action: memoryAction,
139
+ session_id: sessionId,
140
+ };
141
+ if (memoryAction === 'store') {
142
+ const memoryData = this.getNodeParameter('memoryData', i);
143
+ requestBody.data = memoryData;
144
+ }
145
+ const response = await makeN8nAiAgentRequest(this, '/memory', requestBody, false);
146
+ const result = {
147
+ action: memoryAction,
148
+ session_id: sessionId,
149
+ data: response.data || response.memory,
150
+ operation,
151
+ };
152
+ returnData.push({
153
+ json: result,
154
+ pairedItem: { item: i },
155
+ });
156
+ }
157
+ }
158
+ catch (error) {
159
+ if (this.continueOnFail()) {
160
+ const errorMessage = error instanceof Error ? error.message : 'Unknown error';
161
+ returnData.push({
162
+ json: {
163
+ error: errorMessage,
164
+ operation: this.getNodeParameter('operation', i),
165
+ model: this.getNodeParameter('model', i),
166
+ },
167
+ pairedItem: { item: i },
168
+ });
169
+ }
170
+ else {
171
+ throw error;
172
+ }
173
+ }
174
+ }
175
+ return [returnData];
176
+ }
177
+ }
178
+ exports.N8nAiAgent = N8nAiAgent;
179
+ async function makeN8nAiAgentRequest(context, endpoint, requestBody, hasMedia) {
180
+ const credentials = await context.getCredentials('n8nApi');
181
+ const baseUrl = credentials.baseUrl || 'http://localhost:5678';
182
+ const apiKey = credentials.apiKey;
183
+ if (!apiKey) {
184
+ throw new n8n_workflow_1.NodeOperationError(context.getNode(), 'N8N API key is required');
185
+ }
186
+ const options = {
187
+ method: 'POST',
188
+ headers: {
189
+ 'Content-Type': 'application/json',
190
+ 'Authorization': `Bearer ${apiKey}`,
191
+ 'User-Agent': 'N8nAiAgentNode/1.0',
192
+ ...(hasMedia && { 'AI-Vision-Request': 'true' }),
193
+ },
194
+ body: JSON.stringify(requestBody),
195
+ uri: `${baseUrl}/api/v1/ai-agent${endpoint}`,
196
+ json: true,
197
+ };
198
+ try {
199
+ const response = await context.helpers.request(options);
200
+ return response;
201
+ }
202
+ catch (error) {
203
+ const apiError = error;
204
+ if (apiError.statusCode === 401) {
205
+ throw new n8n_workflow_1.NodeOperationError(context.getNode(), 'Invalid N8N API key');
206
+ }
207
+ else if (apiError.statusCode === 404) {
208
+ throw new n8n_workflow_1.NodeOperationError(context.getNode(), 'N8N AI Agent endpoint not found. Make sure AI Agent is enabled in your N8N instance.');
209
+ }
210
+ else {
211
+ throw new n8n_workflow_1.NodeOperationError(context.getNode(), `N8N AI Agent API error: ${apiError.message || String(error)}`);
212
+ }
213
+ }
214
+ }
@@ -0,0 +1,35 @@
1
+ <svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 100 100" width="100" height="100">
2
+ <!-- Background circle -->
3
+ <circle cx="50" cy="50" r="45" fill="#ff6d5a" stroke="#e55a4a" stroke-width="2"/>
4
+
5
+ <!-- N8N Logo style -->
6
+ <g transform="translate(50,50)" fill="white">
7
+ <!-- Brain/Neural network pattern -->
8
+ <circle cx="-15" cy="-15" r="3" opacity="0.9"/>
9
+ <circle cx="0" cy="-20" r="3" opacity="0.9"/>
10
+ <circle cx="15" cy="-15" r="3" opacity="0.9"/>
11
+ <circle cx="-20" cy="0" r="3" opacity="0.9"/>
12
+ <circle cx="0" cy="0" r="4" opacity="1"/>
13
+ <circle cx="20" cy="0" r="3" opacity="0.9"/>
14
+ <circle cx="-15" cy="15" r="3" opacity="0.9"/>
15
+ <circle cx="0" cy="20" r="3" opacity="0.9"/>
16
+ <circle cx="15" cy="15" r="3" opacity="0.9"/>
17
+
18
+ <!-- Connections -->
19
+ <line x1="-15" y1="-15" x2="0" y2="-20" stroke="white" stroke-width="1" opacity="0.6"/>
20
+ <line x1="0" y1="-20" x2="15" y2="-15" stroke="white" stroke-width="1" opacity="0.6"/>
21
+ <line x1="-20" y1="0" x2="-15" y2="-15" stroke="white" stroke-width="1" opacity="0.6"/>
22
+ <line x1="-20" y1="0" x2="0" y2="0" stroke="white" stroke-width="2" opacity="0.8"/>
23
+ <line x1="0" y1="0" x2="20" y2="0" stroke="white" stroke-width="2" opacity="0.8"/>
24
+ <line x1="20" y1="0" x2="15" y2="15" stroke="white" stroke-width="1" opacity="0.6"/>
25
+ <line x1="0" y1="0" x2="0" y2="20" stroke="white" stroke-width="2" opacity="0.8"/>
26
+ <line x1="-15" y1="15" x2="0" y2="20" stroke="white" stroke-width="1" opacity="0.6"/>
27
+ <line x1="0" y1="20" x2="15" y2="15" stroke="white" stroke-width="1" opacity="0.6"/>
28
+
29
+ <!-- AI indicator -->
30
+ <text x="0" y="35" text-anchor="middle" font-family="Arial, sans-serif" font-size="12" font-weight="bold" fill="white">AI</text>
31
+ </g>
32
+
33
+ <!-- N8N text at bottom -->
34
+ <text x="50" y="85" text-anchor="middle" font-family="Arial, sans-serif" font-size="10" font-weight="bold" fill="white">N8N</text>
35
+ </svg>
@@ -0,0 +1,2 @@
1
+ import { INodeProperties } from 'n8n-workflow';
2
+ export declare const nodeProperties: INodeProperties[];
@@ -0,0 +1,432 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.nodeProperties = void 0;
4
+ exports.nodeProperties = [
5
+ {
6
+ displayName: 'Operation',
7
+ name: 'operation',
8
+ type: 'options',
9
+ noDataExpression: true,
10
+ options: [
11
+ {
12
+ name: 'Chat',
13
+ value: 'chat',
14
+ description: 'Send a chat message to N8N AI Agent',
15
+ action: 'Send a chat message',
16
+ },
17
+ {
18
+ name: 'Use Tools',
19
+ value: 'tools',
20
+ description: 'Execute specific tools through AI Agent',
21
+ action: 'Execute tools',
22
+ },
23
+ {
24
+ name: 'Memory',
25
+ value: 'memory',
26
+ description: 'Manage AI Agent memory/context',
27
+ action: 'Manage memory',
28
+ },
29
+ ],
30
+ default: 'chat',
31
+ },
32
+ {
33
+ displayName: 'AI Model',
34
+ name: 'model',
35
+ type: 'options',
36
+ options: [
37
+ {
38
+ name: 'GPT-4 Turbo',
39
+ value: 'gpt-4-turbo',
40
+ description: 'OpenAI GPT-4 Turbo - Latest and most capable',
41
+ },
42
+ {
43
+ name: 'GPT-4',
44
+ value: 'gpt-4',
45
+ description: 'OpenAI GPT-4 - High quality reasoning',
46
+ },
47
+ {
48
+ name: 'GPT-3.5 Turbo',
49
+ value: 'gpt-3.5-turbo',
50
+ description: 'OpenAI GPT-3.5 Turbo - Fast and efficient',
51
+ },
52
+ {
53
+ name: 'Claude 3 Opus',
54
+ value: 'claude-3-opus',
55
+ description: 'Anthropic Claude 3 Opus - Superior reasoning',
56
+ },
57
+ {
58
+ name: 'Claude 3 Sonnet',
59
+ value: 'claude-3-sonnet',
60
+ description: 'Anthropic Claude 3 Sonnet - Balanced performance',
61
+ },
62
+ {
63
+ name: 'Claude 3 Haiku',
64
+ value: 'claude-3-haiku',
65
+ description: 'Anthropic Claude 3 Haiku - Fast responses',
66
+ },
67
+ {
68
+ name: 'Gemini Pro',
69
+ value: 'gemini-pro',
70
+ description: 'Google Gemini Pro - Multimodal capabilities',
71
+ },
72
+ ],
73
+ default: 'gpt-4-turbo',
74
+ description: 'Select the AI model to use',
75
+ },
76
+ {
77
+ displayName: 'Message',
78
+ name: 'message',
79
+ type: 'string',
80
+ typeOptions: {
81
+ rows: 3,
82
+ },
83
+ displayOptions: {
84
+ show: {
85
+ operation: ['chat'],
86
+ },
87
+ },
88
+ default: '',
89
+ placeholder: 'Enter your message here...',
90
+ description: 'The message to send to the AI Agent',
91
+ required: true,
92
+ },
93
+ {
94
+ displayName: 'System Message',
95
+ name: 'systemMessage',
96
+ type: 'string',
97
+ typeOptions: {
98
+ rows: 2,
99
+ },
100
+ displayOptions: {
101
+ show: {
102
+ operation: ['chat'],
103
+ },
104
+ },
105
+ default: '',
106
+ placeholder: 'You are a helpful AI assistant...',
107
+ description: 'System prompt to set the AI behavior and context',
108
+ },
109
+ {
110
+ displayName: 'Include Media',
111
+ name: 'includeMedia',
112
+ type: 'boolean',
113
+ displayOptions: {
114
+ show: {
115
+ operation: ['chat'],
116
+ },
117
+ },
118
+ default: false,
119
+ description: 'Whether to include images in the chat message',
120
+ },
121
+ {
122
+ displayName: 'Media Source',
123
+ name: 'mediaSource',
124
+ type: 'options',
125
+ options: [
126
+ {
127
+ name: 'Upload File',
128
+ value: 'manual',
129
+ description: 'Upload an image file directly',
130
+ },
131
+ {
132
+ name: 'From URL',
133
+ value: 'url',
134
+ description: 'Use an image from URL',
135
+ },
136
+ {
137
+ name: 'From Binary Data',
138
+ value: 'binary',
139
+ description: 'Use image from previous node binary data',
140
+ },
141
+ ],
142
+ displayOptions: {
143
+ show: {
144
+ operation: ['chat'],
145
+ includeMedia: [true],
146
+ },
147
+ },
148
+ default: 'manual',
149
+ description: 'Source of the media file',
150
+ },
151
+ {
152
+ displayName: 'Image File',
153
+ name: 'mediaFile',
154
+ type: 'string',
155
+ displayOptions: {
156
+ show: {
157
+ operation: ['chat'],
158
+ includeMedia: [true],
159
+ mediaSource: ['manual'],
160
+ },
161
+ },
162
+ default: '',
163
+ placeholder: 'Paste base64 image data...',
164
+ description: 'Base64 encoded image data',
165
+ },
166
+ {
167
+ displayName: 'Image URL',
168
+ name: 'mediaUrl',
169
+ type: 'string',
170
+ displayOptions: {
171
+ show: {
172
+ operation: ['chat'],
173
+ includeMedia: [true],
174
+ mediaSource: ['url'],
175
+ },
176
+ },
177
+ default: '',
178
+ placeholder: 'https://example.com/image.jpg',
179
+ description: 'URL of the image to analyze',
180
+ },
181
+ {
182
+ displayName: 'Binary Property',
183
+ name: 'mediaProperty',
184
+ type: 'string',
185
+ displayOptions: {
186
+ show: {
187
+ operation: ['chat'],
188
+ includeMedia: [true],
189
+ mediaSource: ['binary'],
190
+ },
191
+ },
192
+ default: 'data',
193
+ placeholder: 'data',
194
+ description: 'Name of the binary property containing the image',
195
+ },
196
+ {
197
+ displayName: 'Include Conversation History',
198
+ name: 'includeHistory',
199
+ type: 'boolean',
200
+ displayOptions: {
201
+ show: {
202
+ operation: ['chat'],
203
+ },
204
+ },
205
+ default: false,
206
+ description: 'Include previous messages for context',
207
+ },
208
+ {
209
+ displayName: 'Conversation History',
210
+ name: 'conversationHistory',
211
+ type: 'json',
212
+ typeOptions: {
213
+ rows: 4,
214
+ },
215
+ displayOptions: {
216
+ show: {
217
+ operation: ['chat'],
218
+ includeHistory: [true],
219
+ },
220
+ },
221
+ default: '[]',
222
+ placeholder: '[{"role": "user", "content": "Hello"}, {"role": "assistant", "content": "Hi there!"}]',
223
+ description: 'Previous messages in OpenAI chat format',
224
+ },
225
+ {
226
+ displayName: 'Enable Tools',
227
+ name: 'enableTools',
228
+ type: 'boolean',
229
+ displayOptions: {
230
+ show: {
231
+ operation: ['chat'],
232
+ },
233
+ },
234
+ default: false,
235
+ description: 'Allow AI to use tools and function calling',
236
+ },
237
+ {
238
+ displayName: 'Tools Configuration',
239
+ name: 'toolsConfig',
240
+ type: 'json',
241
+ typeOptions: {
242
+ rows: 6,
243
+ },
244
+ displayOptions: {
245
+ show: {
246
+ operation: ['chat'],
247
+ enableTools: [true],
248
+ },
249
+ },
250
+ default: '{"tools": []}',
251
+ placeholder: '{"tools": [{"type": "function", "function": {"name": "get_weather", "description": "Get weather info"}}]}',
252
+ description: 'Tools available to the AI Agent',
253
+ },
254
+ {
255
+ displayName: 'Tool Name',
256
+ name: 'toolName',
257
+ type: 'string',
258
+ displayOptions: {
259
+ show: {
260
+ operation: ['tools'],
261
+ },
262
+ },
263
+ default: '',
264
+ placeholder: 'get_weather',
265
+ description: 'Name of the tool to execute',
266
+ required: true,
267
+ },
268
+ {
269
+ displayName: 'Tool Arguments',
270
+ name: 'toolArguments',
271
+ type: 'json',
272
+ typeOptions: {
273
+ rows: 3,
274
+ },
275
+ displayOptions: {
276
+ show: {
277
+ operation: ['tools'],
278
+ },
279
+ },
280
+ default: '{}',
281
+ placeholder: '{"location": "São Paulo", "units": "celsius"}',
282
+ description: 'Arguments to pass to the tool',
283
+ },
284
+ {
285
+ displayName: 'Context',
286
+ name: 'context',
287
+ type: 'string',
288
+ typeOptions: {
289
+ rows: 2,
290
+ },
291
+ displayOptions: {
292
+ show: {
293
+ operation: ['tools'],
294
+ },
295
+ },
296
+ default: '',
297
+ placeholder: 'User is asking about weather in their city...',
298
+ description: 'Context for tool execution',
299
+ },
300
+ {
301
+ displayName: 'Memory Action',
302
+ name: 'memoryAction',
303
+ type: 'options',
304
+ options: [
305
+ {
306
+ name: 'Store',
307
+ value: 'store',
308
+ description: 'Store data in AI Agent memory',
309
+ },
310
+ {
311
+ name: 'Retrieve',
312
+ value: 'retrieve',
313
+ description: 'Retrieve data from AI Agent memory',
314
+ },
315
+ {
316
+ name: 'Clear',
317
+ value: 'clear',
318
+ description: 'Clear AI Agent memory',
319
+ },
320
+ ],
321
+ displayOptions: {
322
+ show: {
323
+ operation: ['memory'],
324
+ },
325
+ },
326
+ default: 'retrieve',
327
+ description: 'Action to perform on AI Agent memory',
328
+ },
329
+ {
330
+ displayName: 'Session ID',
331
+ name: 'sessionId',
332
+ type: 'string',
333
+ displayOptions: {
334
+ show: {
335
+ operation: ['memory'],
336
+ },
337
+ },
338
+ default: '',
339
+ placeholder: 'user-123',
340
+ description: 'Unique session identifier for memory isolation',
341
+ },
342
+ {
343
+ displayName: 'Memory Data',
344
+ name: 'memoryData',
345
+ type: 'json',
346
+ typeOptions: {
347
+ rows: 3,
348
+ },
349
+ displayOptions: {
350
+ show: {
351
+ operation: ['memory'],
352
+ memoryAction: ['store'],
353
+ },
354
+ },
355
+ default: '{}',
356
+ placeholder: '{"user_preferences": {"language": "pt-BR", "timezone": "America/Sao_Paulo"}}',
357
+ description: 'Data to store in memory',
358
+ },
359
+ {
360
+ displayName: 'Advanced Options',
361
+ name: 'advancedOptions',
362
+ type: 'collection',
363
+ placeholder: 'Add Option',
364
+ default: {},
365
+ displayOptions: {
366
+ show: {
367
+ operation: ['chat'],
368
+ },
369
+ },
370
+ options: [
371
+ {
372
+ displayName: 'Max Tokens',
373
+ name: 'maxTokens',
374
+ type: 'number',
375
+ default: 1000,
376
+ description: 'Maximum number of tokens to generate (0 for model default)',
377
+ typeOptions: {
378
+ minValue: 0,
379
+ maxValue: 4000,
380
+ },
381
+ },
382
+ {
383
+ displayName: 'Temperature',
384
+ name: 'temperature',
385
+ type: 'number',
386
+ default: 0.7,
387
+ description: 'Controls randomness (0.0 = deterministic, 1.0 = very random)',
388
+ typeOptions: {
389
+ minValue: 0,
390
+ maxValue: 1,
391
+ numberPrecision: 2,
392
+ },
393
+ },
394
+ {
395
+ displayName: 'Top P',
396
+ name: 'topP',
397
+ type: 'number',
398
+ default: 1,
399
+ description: 'Controls diversity via nucleus sampling',
400
+ typeOptions: {
401
+ minValue: 0.01,
402
+ maxValue: 1,
403
+ numberPrecision: 2,
404
+ },
405
+ },
406
+ {
407
+ displayName: 'Frequency Penalty',
408
+ name: 'frequencyPenalty',
409
+ type: 'number',
410
+ default: 0,
411
+ description: 'Reduces repetition of words',
412
+ typeOptions: {
413
+ minValue: -2,
414
+ maxValue: 2,
415
+ numberPrecision: 2,
416
+ },
417
+ },
418
+ {
419
+ displayName: 'Presence Penalty',
420
+ name: 'presencePenalty',
421
+ type: 'number',
422
+ default: 0,
423
+ description: 'Encourages new topics',
424
+ typeOptions: {
425
+ minValue: -2,
426
+ maxValue: 2,
427
+ numberPrecision: 2,
428
+ },
429
+ },
430
+ ],
431
+ },
432
+ ];
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "n8n-nodes-github-copilot",
3
- "version": "3.2.6",
4
- "description": "n8n community node for GitHub Copilot with CLI integration and official Chat API access to GPT-5, Claude, Gemini and more using your existing Copilot credits",
3
+ "version": "3.3.0",
4
+ "description": "n8n community node for GitHub Copilot and N8N AI Agent with CLI integration and official Chat API access to GPT-5, Claude, Gemini and more using your existing Copilot credits",
5
5
  "license": "MIT",
6
6
  "homepage": "https://github.com/sufficit/n8n-nodes-github-copilot",
7
7
  "author": {
@@ -32,13 +32,15 @@
32
32
  ],
33
33
  "nodes": [
34
34
  "dist/nodes/GitHubCopilot/GitHubCopilot.node.js",
35
- "dist/nodes/GitHubCopilotChatAPI/GitHubCopilotChatAPI.node.js"
35
+ "dist/nodes/GitHubCopilotChatAPI/GitHubCopilotChatAPI.node.js",
36
+ "dist/nodes/N8nAiAgent/N8nAiAgent.node.js"
36
37
  ]
37
38
  },
38
39
  "keywords": [
39
40
  "n8n-community-node-package",
40
41
  "github",
41
42
  "copilot",
43
+ "n8n-ai-agent",
42
44
  "ai",
43
45
  "gpt-5",
44
46
  "claude",