@hazeljs/kafka 0.2.0-beta.54 → 0.2.0-beta.56

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=kafka-consumer.decorator.test.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"kafka-consumer.decorator.test.d.ts","sourceRoot":"","sources":["../../src/__tests__/kafka-consumer.decorator.test.ts"],"names":[],"mappings":""}
@@ -0,0 +1,100 @@
1
+ "use strict";
2
+ var __decorate = (this && this.__decorate) || function (decorators, target, key, desc) {
3
+ var c = arguments.length, r = c < 3 ? target : desc === null ? desc = Object.getOwnPropertyDescriptor(target, key) : desc, d;
4
+ if (typeof Reflect === "object" && typeof Reflect.decorate === "function") r = Reflect.decorate(decorators, target, key, desc);
5
+ else for (var i = decorators.length - 1; i >= 0; i--) if (d = decorators[i]) r = (c < 3 ? d(r) : c > 3 ? d(target, key, r) : d(target, key)) || r;
6
+ return c > 3 && r && Object.defineProperty(target, key, r), r;
7
+ };
8
+ var __metadata = (this && this.__metadata) || function (k, v) {
9
+ if (typeof Reflect === "object" && typeof Reflect.metadata === "function") return Reflect.metadata(k, v);
10
+ };
11
+ Object.defineProperty(exports, "__esModule", { value: true });
12
+ const kafka_consumer_decorator_1 = require("../decorators/kafka-consumer.decorator");
13
+ const kafka_subscribe_decorator_1 = require("../decorators/kafka-subscribe.decorator");
14
+ describe('KafkaConsumer decorator', () => {
15
+ it('should set consumer metadata on class', () => {
16
+ let TestConsumer = class TestConsumer {
17
+ };
18
+ TestConsumer = __decorate([
19
+ (0, kafka_consumer_decorator_1.KafkaConsumer)({ groupId: 'test-group' })
20
+ ], TestConsumer);
21
+ const metadata = (0, kafka_consumer_decorator_1.getKafkaConsumerMetadata)(TestConsumer.prototype);
22
+ expect(metadata).toBeDefined();
23
+ expect(metadata?.groupId).toBe('test-group');
24
+ expect(metadata?.sessionTimeout).toBe(30000);
25
+ });
26
+ it('should set custom consumer options', () => {
27
+ let CustomConsumer = class CustomConsumer {
28
+ };
29
+ CustomConsumer = __decorate([
30
+ (0, kafka_consumer_decorator_1.KafkaConsumer)({
31
+ groupId: 'custom-group',
32
+ sessionTimeout: 60000,
33
+ heartbeatInterval: 5000,
34
+ })
35
+ ], CustomConsumer);
36
+ const metadata = (0, kafka_consumer_decorator_1.getKafkaConsumerMetadata)(CustomConsumer.prototype);
37
+ expect(metadata?.groupId).toBe('custom-group');
38
+ expect(metadata?.sessionTimeout).toBe(60000);
39
+ expect(metadata?.heartbeatInterval).toBe(5000);
40
+ });
41
+ it('should return true for Kafka consumer class', () => {
42
+ let ConsumerClass = class ConsumerClass {
43
+ };
44
+ ConsumerClass = __decorate([
45
+ (0, kafka_consumer_decorator_1.KafkaConsumer)({ groupId: 'test' })
46
+ ], ConsumerClass);
47
+ expect((0, kafka_consumer_decorator_1.isKafkaConsumer)(ConsumerClass.prototype)).toBe(true);
48
+ });
49
+ it('should return false for non-consumer class', () => {
50
+ class RegularClass {
51
+ }
52
+ expect((0, kafka_consumer_decorator_1.isKafkaConsumer)(RegularClass.prototype)).toBe(false);
53
+ });
54
+ });
55
+ describe('KafkaSubscribe decorator', () => {
56
+ it('should set subscribe metadata on method', () => {
57
+ let TestConsumer = class TestConsumer {
58
+ handleOrders() { }
59
+ };
60
+ __decorate([
61
+ (0, kafka_subscribe_decorator_1.KafkaSubscribe)('orders'),
62
+ __metadata("design:type", Function),
63
+ __metadata("design:paramtypes", []),
64
+ __metadata("design:returntype", void 0)
65
+ ], TestConsumer.prototype, "handleOrders", null);
66
+ TestConsumer = __decorate([
67
+ (0, kafka_consumer_decorator_1.KafkaConsumer)({ groupId: 'test-group' })
68
+ ], TestConsumer);
69
+ const metadata = (0, kafka_subscribe_decorator_1.getKafkaSubscribeMetadata)(TestConsumer.prototype);
70
+ expect(metadata).toHaveLength(1);
71
+ expect(metadata[0].topic).toBe('orders');
72
+ expect(metadata[0].methodName).toBe('handleOrders');
73
+ });
74
+ it('should support multiple topic subscriptions', () => {
75
+ let MultiConsumer = class MultiConsumer {
76
+ handleA() { }
77
+ handleB() { }
78
+ };
79
+ __decorate([
80
+ (0, kafka_subscribe_decorator_1.KafkaSubscribe)('topic-a'),
81
+ __metadata("design:type", Function),
82
+ __metadata("design:paramtypes", []),
83
+ __metadata("design:returntype", void 0)
84
+ ], MultiConsumer.prototype, "handleA", null);
85
+ __decorate([
86
+ (0, kafka_subscribe_decorator_1.KafkaSubscribe)('topic-b', { fromBeginning: true }),
87
+ __metadata("design:type", Function),
88
+ __metadata("design:paramtypes", []),
89
+ __metadata("design:returntype", void 0)
90
+ ], MultiConsumer.prototype, "handleB", null);
91
+ MultiConsumer = __decorate([
92
+ (0, kafka_consumer_decorator_1.KafkaConsumer)({ groupId: 'test-group' })
93
+ ], MultiConsumer);
94
+ const metadata = (0, kafka_subscribe_decorator_1.getKafkaSubscribeMetadata)(MultiConsumer.prototype);
95
+ expect(metadata).toHaveLength(2);
96
+ expect(metadata[0]).toEqual({ topic: 'topic-a', methodName: 'handleA', options: {} });
97
+ expect(metadata[1].topic).toBe('topic-b');
98
+ expect(metadata[1].options?.fromBeginning).toBe(true);
99
+ });
100
+ });
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=kafka-consumer.service.test.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"kafka-consumer.service.test.d.ts","sourceRoot":"","sources":["../../src/__tests__/kafka-consumer.service.test.ts"],"names":[],"mappings":""}
@@ -0,0 +1,244 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ const kafka_consumer_service_1 = require("../kafka-consumer.service");
4
+ const kafka_consumer_decorator_1 = require("../decorators/kafka-consumer.decorator");
5
+ const kafka_subscribe_decorator_1 = require("../decorators/kafka-subscribe.decorator");
6
+ // Mock metadata reader functions to control what the service sees
7
+ jest.mock('../decorators/kafka-consumer.decorator', () => ({
8
+ getKafkaConsumerMetadata: jest.fn(),
9
+ KAFKA_CONSUMER_METADATA_KEY: Symbol('kafka:consumer'),
10
+ }));
11
+ jest.mock('../decorators/kafka-subscribe.decorator', () => ({
12
+ getKafkaSubscribeMetadata: jest.fn(),
13
+ KAFKA_SUBSCRIBE_METADATA_KEY: Symbol('kafka:subscribe'),
14
+ }));
15
+ // Mock kafkajs
16
+ const mockConnect = jest.fn().mockResolvedValue(undefined);
17
+ const mockDisconnect = jest.fn().mockResolvedValue(undefined);
18
+ const mockSubscribe = jest.fn().mockResolvedValue(undefined);
19
+ let capturedEachMessage;
20
+ const mockRun = jest
21
+ .fn()
22
+ .mockImplementation(async ({ eachMessage }) => {
23
+ capturedEachMessage = eachMessage;
24
+ });
25
+ jest.mock('kafkajs', () => ({ Kafka: jest.fn() }));
26
+ const mockConsumer = {
27
+ connect: mockConnect,
28
+ disconnect: mockDisconnect,
29
+ subscribe: mockSubscribe,
30
+ run: mockRun,
31
+ };
32
+ const mockKafka = {
33
+ consumer: jest.fn().mockReturnValue(mockConsumer),
34
+ };
35
+ const mockGetConsumerMetadata = kafka_consumer_decorator_1.getKafkaConsumerMetadata;
36
+ const mockGetSubscribeMetadata = kafka_subscribe_decorator_1.getKafkaSubscribeMetadata;
37
+ describe('KafkaConsumerService', () => {
38
+ let service;
39
+ beforeEach(() => {
40
+ jest.clearAllMocks();
41
+ capturedEachMessage = undefined;
42
+ mockConnect.mockResolvedValue(undefined);
43
+ mockDisconnect.mockResolvedValue(undefined);
44
+ mockSubscribe.mockResolvedValue(undefined);
45
+ mockRun.mockImplementation(async ({ eachMessage }) => {
46
+ capturedEachMessage = eachMessage;
47
+ });
48
+ mockKafka.consumer.mockReturnValue(mockConsumer);
49
+ service = new kafka_consumer_service_1.KafkaConsumerService(mockKafka);
50
+ });
51
+ describe('getConsumerCount()', () => {
52
+ it('returns 0 initially', () => {
53
+ expect(service.getConsumerCount()).toBe(0);
54
+ });
55
+ });
56
+ describe('registerFromProvider()', () => {
57
+ it('warns and returns when provider has no @KafkaConsumer', async () => {
58
+ mockGetConsumerMetadata.mockReturnValue(undefined);
59
+ mockGetSubscribeMetadata.mockReturnValue([]);
60
+ class NoMetadataProvider {
61
+ }
62
+ await service.registerFromProvider(new NoMetadataProvider());
63
+ expect(mockConnect).not.toHaveBeenCalled();
64
+ });
65
+ it('warns and returns when provider has @KafkaConsumer but empty @KafkaSubscribe', async () => {
66
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'test-group' });
67
+ mockGetSubscribeMetadata.mockReturnValue([]);
68
+ class NoSubscribeProvider {
69
+ }
70
+ await service.registerFromProvider(new NoSubscribeProvider());
71
+ expect(mockConnect).not.toHaveBeenCalled();
72
+ });
73
+ it('registers and starts a consumer', async () => {
74
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'test-group' });
75
+ mockGetSubscribeMetadata.mockReturnValue([
76
+ { topic: 'test-topic', methodName: 'handleMessage', options: {} },
77
+ ]);
78
+ class TestProvider {
79
+ async handleMessage(_payload) { }
80
+ }
81
+ await service.registerFromProvider(new TestProvider());
82
+ expect(mockConnect).toHaveBeenCalledTimes(1);
83
+ expect(mockSubscribe).toHaveBeenCalledWith({ topics: ['test-topic'], fromBeginning: false });
84
+ expect(mockRun).toHaveBeenCalledTimes(1);
85
+ expect(service.getConsumerCount()).toBe(1);
86
+ });
87
+ it('subscribes with fromBeginning=true when specified', async () => {
88
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'replay-group' });
89
+ mockGetSubscribeMetadata.mockReturnValue([
90
+ { topic: 'replay-topic', methodName: 'handle', options: { fromBeginning: true } },
91
+ ]);
92
+ class ReplayProvider {
93
+ async handle(_payload) { }
94
+ }
95
+ await service.registerFromProvider(new ReplayProvider());
96
+ expect(mockSubscribe).toHaveBeenCalledWith({ topics: ['replay-topic'], fromBeginning: true });
97
+ });
98
+ it('passes consumer options to kafka.consumer()', async () => {
99
+ mockGetConsumerMetadata.mockReturnValue({
100
+ groupId: 'my-group',
101
+ sessionTimeout: 10000,
102
+ rebalanceTimeout: 20000,
103
+ heartbeatInterval: 1000,
104
+ maxWaitTimeInMs: 2000,
105
+ });
106
+ mockGetSubscribeMetadata.mockReturnValue([
107
+ { topic: 'topic', methodName: 'handle', options: {} },
108
+ ]);
109
+ class Provider {
110
+ async handle(_payload) { }
111
+ }
112
+ await service.registerFromProvider(new Provider());
113
+ expect(mockKafka.consumer).toHaveBeenCalledWith(expect.objectContaining({ groupId: 'my-group', sessionTimeout: 10000 }));
114
+ });
115
+ it('dispatches eachMessage to the correct handler', async () => {
116
+ const handleSpy = jest.fn().mockResolvedValue(undefined);
117
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'order-group' });
118
+ mockGetSubscribeMetadata.mockReturnValue([
119
+ { topic: 'orders', methodName: 'processOrder', options: {} },
120
+ ]);
121
+ class OrderProvider {
122
+ async processOrder(payload) {
123
+ handleSpy(payload);
124
+ }
125
+ }
126
+ const provider = new OrderProvider();
127
+ await service.registerFromProvider(provider);
128
+ expect(capturedEachMessage).toBeDefined();
129
+ const testPayload = {
130
+ topic: 'orders',
131
+ partition: 0,
132
+ message: {
133
+ key: null,
134
+ value: Buffer.from('{"id":1}'),
135
+ headers: {},
136
+ offset: '0',
137
+ timestamp: '1000',
138
+ },
139
+ heartbeat: jest.fn(),
140
+ pause: jest.fn(),
141
+ };
142
+ await capturedEachMessage(testPayload);
143
+ expect(handleSpy).toHaveBeenCalledWith(testPayload);
144
+ });
145
+ it('ignores messages for unregistered topics', async () => {
146
+ const handleSpy = jest.fn();
147
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'g' });
148
+ mockGetSubscribeMetadata.mockReturnValue([
149
+ { topic: 'known-topic', methodName: 'handle', options: {} },
150
+ ]);
151
+ class Provider {
152
+ async handle() {
153
+ handleSpy();
154
+ }
155
+ }
156
+ await service.registerFromProvider(new Provider());
157
+ await capturedEachMessage({
158
+ topic: 'unknown-topic',
159
+ partition: 0,
160
+ message: { key: null, value: Buffer.from('x'), headers: {}, offset: '0', timestamp: '0' },
161
+ heartbeat: jest.fn(),
162
+ pause: jest.fn(),
163
+ });
164
+ expect(handleSpy).not.toHaveBeenCalled();
165
+ });
166
+ it('handles handler errors gracefully without throwing', async () => {
167
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'g' });
168
+ mockGetSubscribeMetadata.mockReturnValue([
169
+ { topic: 'error-topic', methodName: 'handle', options: {} },
170
+ ]);
171
+ class ErrorProvider {
172
+ async handle() {
173
+ throw new Error('handler blew up');
174
+ }
175
+ }
176
+ await service.registerFromProvider(new ErrorProvider());
177
+ await expect(capturedEachMessage({
178
+ topic: 'error-topic',
179
+ partition: 0,
180
+ message: { key: null, value: Buffer.from('x'), headers: {}, offset: '0', timestamp: '0' },
181
+ heartbeat: jest.fn(),
182
+ pause: jest.fn(),
183
+ })).resolves.toBeUndefined();
184
+ });
185
+ it('logs error when method name is not a function on provider', async () => {
186
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'g' });
187
+ mockGetSubscribeMetadata.mockReturnValue([
188
+ { topic: 'test-topic', methodName: 'nonExistentMethod', options: {} },
189
+ ]);
190
+ class ProviderWithoutMethod {
191
+ }
192
+ await service.registerFromProvider(new ProviderWithoutMethod());
193
+ // Should not throw, just log an error
194
+ await expect(capturedEachMessage({
195
+ topic: 'test-topic',
196
+ partition: 0,
197
+ message: { key: null, value: Buffer.from('x'), headers: {}, offset: '0', timestamp: '0' },
198
+ heartbeat: jest.fn(),
199
+ pause: jest.fn(),
200
+ })).resolves.toBeUndefined();
201
+ });
202
+ it('supports multiple subscriptions on one provider', async () => {
203
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'multi-group' });
204
+ mockGetSubscribeMetadata.mockReturnValue([
205
+ { topic: 'topic-a', methodName: 'handleA', options: {} },
206
+ { topic: 'topic-b', methodName: 'handleB', options: {} },
207
+ ]);
208
+ class MultiProvider {
209
+ async handleA(_p) { }
210
+ async handleB(_p) { }
211
+ }
212
+ await service.registerFromProvider(new MultiProvider());
213
+ expect(mockSubscribe).toHaveBeenCalledTimes(2);
214
+ });
215
+ });
216
+ describe('onModuleDestroy()', () => {
217
+ it('does nothing when no consumers are running', async () => {
218
+ await service.onModuleDestroy();
219
+ expect(mockDisconnect).not.toHaveBeenCalled();
220
+ });
221
+ it('disconnects all running consumers and clears list', async () => {
222
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'g' });
223
+ mockGetSubscribeMetadata.mockReturnValue([{ topic: 't', methodName: 'h', options: {} }]);
224
+ class P {
225
+ async h() { }
226
+ }
227
+ await service.registerFromProvider(new P());
228
+ expect(service.getConsumerCount()).toBe(1);
229
+ await service.onModuleDestroy();
230
+ expect(mockDisconnect).toHaveBeenCalledTimes(1);
231
+ expect(service.getConsumerCount()).toBe(0);
232
+ });
233
+ it('handles disconnect errors gracefully', async () => {
234
+ mockGetConsumerMetadata.mockReturnValue({ groupId: 'g' });
235
+ mockGetSubscribeMetadata.mockReturnValue([{ topic: 't', methodName: 'h', options: {} }]);
236
+ class P {
237
+ async h() { }
238
+ }
239
+ await service.registerFromProvider(new P());
240
+ mockDisconnect.mockRejectedValueOnce(new Error('disconnect failed'));
241
+ await expect(service.onModuleDestroy()).resolves.toBeUndefined();
242
+ });
243
+ });
244
+ });
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=kafka-producer.service.test.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"kafka-producer.service.test.d.ts","sourceRoot":"","sources":["../../src/__tests__/kafka-producer.service.test.ts"],"names":[],"mappings":""}
@@ -0,0 +1,73 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ const kafka_producer_service_1 = require("../kafka-producer.service");
4
+ // Mock kafkajs
5
+ const mockSend = jest.fn().mockResolvedValue(undefined);
6
+ const mockConnect = jest.fn().mockResolvedValue(undefined);
7
+ const mockDisconnect = jest.fn().mockResolvedValue(undefined);
8
+ jest.mock('kafkajs', () => ({
9
+ Kafka: jest.fn().mockImplementation(() => ({
10
+ producer: () => ({
11
+ connect: mockConnect,
12
+ disconnect: mockDisconnect,
13
+ send: mockSend,
14
+ }),
15
+ })),
16
+ }));
17
+ // Simple DI container mock for testing
18
+ const mockKafka = {
19
+ producer: () => ({
20
+ connect: mockConnect,
21
+ disconnect: mockDisconnect,
22
+ send: mockSend,
23
+ }),
24
+ };
25
+ describe('KafkaProducerService', () => {
26
+ let service;
27
+ beforeEach(() => {
28
+ jest.clearAllMocks();
29
+ service = new kafka_producer_service_1.KafkaProducerService(mockKafka);
30
+ });
31
+ describe('send', () => {
32
+ it('should connect and send message', async () => {
33
+ await service.send('test-topic', { value: 'hello' });
34
+ expect(mockConnect).toHaveBeenCalled();
35
+ expect(mockSend).toHaveBeenCalledWith(expect.objectContaining({
36
+ topic: 'test-topic',
37
+ messages: expect.arrayContaining([expect.objectContaining({ value: 'hello' })]),
38
+ }));
39
+ });
40
+ it('should send multiple messages', async () => {
41
+ await service.send('test-topic', [
42
+ { key: '1', value: 'msg1' },
43
+ { key: '2', value: 'msg2' },
44
+ ]);
45
+ expect(mockSend).toHaveBeenCalledWith(expect.objectContaining({
46
+ messages: expect.arrayContaining([
47
+ expect.objectContaining({ key: '1', value: 'msg1' }),
48
+ expect.objectContaining({ key: '2', value: 'msg2' }),
49
+ ]),
50
+ }));
51
+ });
52
+ });
53
+ describe('sendBatch', () => {
54
+ it('should send to multiple topics', async () => {
55
+ await service.sendBatch([
56
+ { topic: 'topic-a', messages: [{ value: 'a' }] },
57
+ { topic: 'topic-b', messages: [{ value: 'b' }] },
58
+ ]);
59
+ expect(mockSend).toHaveBeenCalledTimes(2);
60
+ });
61
+ });
62
+ describe('lifecycle', () => {
63
+ it('should connect on onModuleInit', async () => {
64
+ await service.onModuleInit();
65
+ expect(mockConnect).toHaveBeenCalled();
66
+ });
67
+ it('should disconnect on onModuleDestroy', async () => {
68
+ await service.onModuleInit();
69
+ await service.onModuleDestroy();
70
+ expect(mockDisconnect).toHaveBeenCalled();
71
+ });
72
+ });
73
+ });
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=kafka-stream.processor.test.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"kafka-stream.processor.test.d.ts","sourceRoot":"","sources":["../../src/__tests__/kafka-stream.processor.test.ts"],"names":[],"mappings":""}
@@ -0,0 +1,243 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ const kafka_stream_processor_1 = require("../kafka-stream.processor");
4
+ // Mock kafkajs
5
+ const mockConsumerConnect = jest.fn().mockResolvedValue(undefined);
6
+ const mockConsumerDisconnect = jest.fn().mockResolvedValue(undefined);
7
+ const mockConsumerSubscribe = jest.fn().mockResolvedValue(undefined);
8
+ let capturedEachMessage;
9
+ const mockConsumerRun = jest
10
+ .fn()
11
+ .mockImplementation(async ({ eachMessage }) => {
12
+ capturedEachMessage = eachMessage;
13
+ });
14
+ const mockProducerConnect = jest.fn().mockResolvedValue(undefined);
15
+ const mockProducerDisconnect = jest.fn().mockResolvedValue(undefined);
16
+ const mockProducerSend = jest.fn().mockResolvedValue(undefined);
17
+ jest.mock('kafkajs', () => ({ Kafka: jest.fn() }));
18
+ const mockKafka = {
19
+ consumer: jest.fn().mockReturnValue({
20
+ connect: mockConsumerConnect,
21
+ disconnect: mockConsumerDisconnect,
22
+ subscribe: mockConsumerSubscribe,
23
+ run: mockConsumerRun,
24
+ }),
25
+ producer: jest.fn().mockReturnValue({
26
+ connect: mockProducerConnect,
27
+ disconnect: mockProducerDisconnect,
28
+ send: mockProducerSend,
29
+ }),
30
+ };
31
+ describe('KafkaStreamProcessor', () => {
32
+ let processor;
33
+ beforeEach(() => {
34
+ jest.clearAllMocks();
35
+ capturedEachMessage = undefined;
36
+ mockConsumerConnect.mockResolvedValue(undefined);
37
+ mockConsumerDisconnect.mockResolvedValue(undefined);
38
+ mockConsumerSubscribe.mockResolvedValue(undefined);
39
+ mockConsumerRun.mockImplementation(async ({ eachMessage }) => {
40
+ capturedEachMessage = eachMessage;
41
+ });
42
+ mockProducerConnect.mockResolvedValue(undefined);
43
+ mockProducerDisconnect.mockResolvedValue(undefined);
44
+ mockProducerSend.mockResolvedValue(undefined);
45
+ mockKafka.consumer.mockReturnValue({
46
+ connect: mockConsumerConnect,
47
+ disconnect: mockConsumerDisconnect,
48
+ subscribe: mockConsumerSubscribe,
49
+ run: mockConsumerRun,
50
+ });
51
+ mockKafka.producer.mockReturnValue({
52
+ connect: mockProducerConnect,
53
+ disconnect: mockProducerDisconnect,
54
+ send: mockProducerSend,
55
+ });
56
+ processor = new kafka_stream_processor_1.KafkaStreamProcessor(mockKafka);
57
+ });
58
+ describe('fluent API', () => {
59
+ it('from() returns this (chainable)', () => {
60
+ expect(processor.from('input-topic')).toBe(processor);
61
+ });
62
+ it('from() can be called again to update inputTopic', () => {
63
+ processor.from('first');
64
+ expect(processor.from('second')).toBe(processor);
65
+ });
66
+ it('transform() returns this (chainable)', () => {
67
+ processor.from('input');
68
+ expect(processor.transform(async (msg) => ({ value: msg.value }))).toBe(processor);
69
+ });
70
+ it('to() returns this (chainable)', () => {
71
+ processor.from('input');
72
+ expect(processor.to('output')).toBe(processor);
73
+ });
74
+ it('withGroupId() returns this (chainable)', () => {
75
+ processor.from('input');
76
+ expect(processor.withGroupId('my-group')).toBe(processor);
77
+ });
78
+ it('transform() throws when called before from()', () => {
79
+ expect(() => processor.transform(async (msg) => ({ value: msg.value }))).toThrow('Call from(topic) before transform()');
80
+ });
81
+ it('to() throws when called before from()', () => {
82
+ expect(() => processor.to('output')).toThrow('Call from(topic) before to()');
83
+ });
84
+ it('withGroupId() throws when called before from()', () => {
85
+ expect(() => processor.withGroupId('group')).toThrow('Call from(topic) before withGroupId()');
86
+ });
87
+ });
88
+ describe('isProcessorRunning()', () => {
89
+ it('returns false initially', () => {
90
+ expect(processor.isProcessorRunning()).toBe(false);
91
+ });
92
+ });
93
+ describe('start()', () => {
94
+ it('throws when pipeline is not configured', async () => {
95
+ await expect(processor.start()).rejects.toThrow('Pipeline not configured');
96
+ });
97
+ it('starts the processor and sets isRunning', async () => {
98
+ processor.from('input').to('output');
99
+ await processor.start();
100
+ expect(mockConsumerConnect).toHaveBeenCalled();
101
+ expect(mockProducerConnect).toHaveBeenCalled();
102
+ expect(mockConsumerSubscribe).toHaveBeenCalledWith({
103
+ topics: ['input'],
104
+ fromBeginning: false,
105
+ });
106
+ expect(processor.isProcessorRunning()).toBe(true);
107
+ });
108
+ it('is idempotent (second start() is no-op)', async () => {
109
+ processor.from('input').to('output');
110
+ await processor.start();
111
+ await processor.start();
112
+ expect(mockConsumerConnect).toHaveBeenCalledTimes(1);
113
+ });
114
+ it('uses custom groupId when provided', async () => {
115
+ processor.from('input').to('output').withGroupId('custom-group');
116
+ await processor.start();
117
+ expect(mockKafka.consumer).toHaveBeenCalledWith({ groupId: 'custom-group' });
118
+ });
119
+ it('generates default groupId from topic names', async () => {
120
+ processor.from('in-topic').to('out-topic');
121
+ await processor.start();
122
+ expect(mockKafka.consumer).toHaveBeenCalledWith({
123
+ groupId: 'stream-in-topic-out-topic',
124
+ });
125
+ });
126
+ it('produces transformed message to output topic', async () => {
127
+ processor
128
+ .from('in')
129
+ .transform(async (msg) => ({ value: Buffer.from(`processed:${msg.value?.toString()}`) }))
130
+ .to('out');
131
+ await processor.start();
132
+ await capturedEachMessage({
133
+ message: { key: Buffer.from('key1'), value: Buffer.from('hello'), headers: {} },
134
+ });
135
+ expect(mockProducerSend).toHaveBeenCalledWith(expect.objectContaining({ topic: 'out' }));
136
+ });
137
+ it('skips producing when transform returns null', async () => {
138
+ processor
139
+ .from('in')
140
+ .transform(async () => null)
141
+ .to('out');
142
+ await processor.start();
143
+ await capturedEachMessage({
144
+ message: { key: null, value: Buffer.from('data'), headers: {} },
145
+ });
146
+ expect(mockProducerSend).not.toHaveBeenCalled();
147
+ });
148
+ it('serializes non-buffer/string output to JSON', async () => {
149
+ processor
150
+ .from('in')
151
+ .transform(async () => ({ value: { nested: 'obj' } }))
152
+ .to('out');
153
+ await processor.start();
154
+ await capturedEachMessage({
155
+ message: { key: null, value: Buffer.from('in'), headers: {} },
156
+ });
157
+ const sentMsg = mockProducerSend.mock.calls[0][0].messages[0];
158
+ expect(sentMsg.value).toBe('{"nested":"obj"}');
159
+ });
160
+ it('passes string value through as-is', async () => {
161
+ processor
162
+ .from('in')
163
+ .transform(async () => ({ value: 'plain-text' }))
164
+ .to('out');
165
+ await processor.start();
166
+ await capturedEachMessage({
167
+ message: { key: null, value: Buffer.from('in'), headers: {} },
168
+ });
169
+ const sentMsg = mockProducerSend.mock.calls[0][0].messages[0];
170
+ expect(sentMsg.value).toBe('plain-text');
171
+ });
172
+ it('passes Buffer value through as-is', async () => {
173
+ const buf = Buffer.from('raw');
174
+ processor
175
+ .from('in')
176
+ .transform(async () => ({ value: buf }))
177
+ .to('out');
178
+ await processor.start();
179
+ await capturedEachMessage({
180
+ message: { key: null, value: Buffer.from('in'), headers: {} },
181
+ });
182
+ const sentMsg = mockProducerSend.mock.calls[0][0].messages[0];
183
+ expect(sentMsg.value).toBe(buf);
184
+ });
185
+ it('uses original message value when transform output value is null', async () => {
186
+ processor
187
+ .from('in')
188
+ .transform(async (msg) => ({ value: null, key: msg.key }))
189
+ .to('out');
190
+ await processor.start();
191
+ const origValue = Buffer.from('original');
192
+ await capturedEachMessage({
193
+ message: { key: null, value: origValue, headers: {} },
194
+ });
195
+ const sentMsg = mockProducerSend.mock.calls[0][0].messages[0];
196
+ expect(sentMsg.value).toBe(origValue);
197
+ });
198
+ it('handles transform errors by re-throwing', async () => {
199
+ processor
200
+ .from('in')
201
+ .transform(async () => {
202
+ throw new Error('transform failed');
203
+ })
204
+ .to('out');
205
+ await processor.start();
206
+ await expect(capturedEachMessage({
207
+ message: { key: null, value: Buffer.from('x'), headers: {} },
208
+ })).rejects.toThrow('transform failed');
209
+ });
210
+ it('uses identity transform when no transform() is called', async () => {
211
+ processor.from('in').to('out');
212
+ await processor.start();
213
+ const val = Buffer.from('unchanged');
214
+ await capturedEachMessage({
215
+ message: { key: null, value: val, headers: {} },
216
+ });
217
+ expect(mockProducerSend).toHaveBeenCalled();
218
+ });
219
+ });
220
+ describe('stop()', () => {
221
+ it('does nothing when processor is not running', async () => {
222
+ await processor.stop();
223
+ expect(mockConsumerDisconnect).not.toHaveBeenCalled();
224
+ });
225
+ it('disconnects both consumer and producer', async () => {
226
+ processor.from('in').to('out');
227
+ await processor.start();
228
+ await processor.stop();
229
+ expect(mockConsumerDisconnect).toHaveBeenCalled();
230
+ expect(mockProducerDisconnect).toHaveBeenCalled();
231
+ expect(processor.isProcessorRunning()).toBe(false);
232
+ });
233
+ it('resets pipelineConfig after stop', async () => {
234
+ processor.from('in').to('out');
235
+ await processor.start();
236
+ await processor.stop();
237
+ // Can start a new pipeline after stop
238
+ processor.from('new-in').to('new-out');
239
+ await processor.start();
240
+ expect(processor.isProcessorRunning()).toBe(true);
241
+ });
242
+ });
243
+ });
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=kafka.module.test.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"kafka.module.test.d.ts","sourceRoot":"","sources":["../../src/__tests__/kafka.module.test.ts"],"names":[],"mappings":""}
@@ -0,0 +1,41 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ const kafka_module_1 = require("../kafka.module");
4
+ const core_1 = require("@hazeljs/core");
5
+ const kafka_producer_service_1 = require("../kafka-producer.service");
6
+ describe('KafkaModule', () => {
7
+ describe('forRoot', () => {
8
+ it('should return KafkaModule and register client', () => {
9
+ const result = kafka_module_1.KafkaModule.forRoot({
10
+ clientId: 'test-app',
11
+ brokers: ['localhost:9092'],
12
+ });
13
+ expect(result).toBe(kafka_module_1.KafkaModule);
14
+ const container = core_1.Container.getInstance();
15
+ const client = container.resolve(kafka_producer_service_1.KAFKA_CLIENT_TOKEN);
16
+ expect(client).toBeDefined();
17
+ expect(client?.producer).toBeDefined();
18
+ });
19
+ it('should use default options', () => {
20
+ kafka_module_1.KafkaModule.forRoot({});
21
+ const container = core_1.Container.getInstance();
22
+ const client = container.resolve(kafka_producer_service_1.KAFKA_CLIENT_TOKEN);
23
+ expect(client).toBeDefined();
24
+ });
25
+ });
26
+ describe('forRootAsync', () => {
27
+ it('should return KafkaModule and register client', async () => {
28
+ const result = await kafka_module_1.KafkaModule.forRootAsync({
29
+ useFactory: () => Promise.resolve({
30
+ clientId: 'async-app',
31
+ brokers: ['kafka:9092'],
32
+ }),
33
+ inject: [],
34
+ });
35
+ expect(result).toBe(kafka_module_1.KafkaModule);
36
+ const container = core_1.Container.getInstance();
37
+ const client = container.resolve(kafka_producer_service_1.KAFKA_CLIENT_TOKEN);
38
+ expect(client).toBeDefined();
39
+ });
40
+ });
41
+ });
@@ -112,7 +112,7 @@ let KafkaConsumerService = class KafkaConsumerService {
112
112
  };
113
113
  exports.KafkaConsumerService = KafkaConsumerService;
114
114
  exports.KafkaConsumerService = KafkaConsumerService = __decorate([
115
- (0, core_1.Injectable)(),
115
+ (0, core_1.Service)(),
116
116
  __param(0, (0, core_1.Inject)(exports.KAFKA_CLIENT_TOKEN)),
117
117
  __metadata("design:paramtypes", [kafkajs_1.Kafka])
118
118
  ], KafkaConsumerService);
@@ -101,7 +101,7 @@ let KafkaProducerService = class KafkaProducerService {
101
101
  };
102
102
  exports.KafkaProducerService = KafkaProducerService;
103
103
  exports.KafkaProducerService = KafkaProducerService = __decorate([
104
- (0, core_1.Injectable)(),
104
+ (0, core_1.Service)(),
105
105
  __param(0, (0, core_1.Inject)(exports.KAFKA_CLIENT_TOKEN)),
106
106
  __metadata("design:paramtypes", [kafkajs_1.Kafka])
107
107
  ], KafkaProducerService);
@@ -162,7 +162,7 @@ let KafkaStreamProcessor = class KafkaStreamProcessor {
162
162
  };
163
163
  exports.KafkaStreamProcessor = KafkaStreamProcessor;
164
164
  exports.KafkaStreamProcessor = KafkaStreamProcessor = __decorate([
165
- (0, core_1.Injectable)(),
165
+ (0, core_1.Service)(),
166
166
  __param(0, (0, core_1.Inject)(exports.KAFKA_CLIENT_TOKEN)),
167
167
  __metadata("design:paramtypes", [kafkajs_1.Kafka])
168
168
  ], KafkaStreamProcessor);
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@hazeljs/kafka",
3
- "version": "0.2.0-beta.54",
3
+ "version": "0.2.0-beta.56",
4
4
  "description": "Kafka module for HazelJS framework - produce, consume, and stream processing",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
@@ -50,5 +50,5 @@
50
50
  "peerDependencies": {
51
51
  "@hazeljs/core": ">=0.2.0-beta.0"
52
52
  },
53
- "gitHead": "c593ce33447cdc62d7bd2386cc2db47840292fcb"
53
+ "gitHead": "c2737e90974458a8438eee623726f0a453b66b8b"
54
54
  }