kafka-console 1.2.58 → 1.2.59

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -54,9 +54,9 @@ const Fs = __importStar(require("fs"));
54
54
  const kafka_1 = require("../utils/kafka");
55
55
  const formatters_1 = require("../utils/formatters");
56
56
  function consume(topic, opts, { parent }) {
57
- var e_1, _a;
57
+ var _a, e_1, _b, _c;
58
58
  return __awaiter(this, void 0, void 0, function* () {
59
- const _b = Object.assign(Object.assign({}, parent.opts()), opts), { group, format, fromBeginning, count, skip, filename, brokers, logLevel, timeout, ssl } = _b, saslOptions = __rest(_b, ["group", "format", "fromBeginning", "count", "skip", "filename", "brokers", "logLevel", "timeout", "ssl"]);
59
+ const _d = Object.assign(Object.assign({}, parent.opts()), opts), { group, format, fromBeginning, count, skip, filename, brokers, logLevel, timeout, ssl } = _d, saslOptions = __rest(_d, ["group", "format", "fromBeginning", "count", "skip", "filename", "brokers", "logLevel", "timeout", "ssl"]);
60
60
  const sasl = (0, kafka_1.getSASL)(saslOptions);
61
61
  const client = (0, kafka_1.createClient)(brokers, ssl, sasl, logLevel);
62
62
  const output = filename ? Fs.createWriteStream(filename) : process.stdout;
@@ -69,19 +69,26 @@ function consume(topic, opts, { parent }) {
69
69
  });
70
70
  const formatter = (0, formatters_1.getFormatter)(format);
71
71
  try {
72
- for (var consumer_1 = __asyncValues(consumer), consumer_1_1; consumer_1_1 = yield consumer_1.next(), !consumer_1_1.done;) {
73
- let { message: { headers, key, value } } = consumer_1_1.value;
74
- const parsedHeaders = Object.entries(headers).reduce((result, [key, value]) => {
75
- return Object.assign(Object.assign({}, result), { [key]: value.toString() });
76
- }, {});
77
- const message = { headers: parsedHeaders, key: key === null || key === void 0 ? void 0 : key.toString(), value: yield formatter.decode(value) };
78
- output.write(JSON.stringify(message, null, ' ') + '\n');
72
+ for (var _e = true, consumer_1 = __asyncValues(consumer), consumer_1_1; consumer_1_1 = yield consumer_1.next(), _a = consumer_1_1.done, !_a;) {
73
+ _c = consumer_1_1.value;
74
+ _e = false;
75
+ try {
76
+ let { message: { headers, key, value } } = _c;
77
+ const parsedHeaders = Object.entries(headers).reduce((result, [key, value]) => {
78
+ return Object.assign(Object.assign({}, result), { [key]: value.toString() });
79
+ }, {});
80
+ const message = { headers: parsedHeaders, key: key === null || key === void 0 ? void 0 : key.toString(), value: yield formatter.decode(value) };
81
+ output.write(JSON.stringify(message, null, ' ') + '\n');
82
+ }
83
+ finally {
84
+ _e = true;
85
+ }
79
86
  }
80
87
  }
81
88
  catch (e_1_1) { e_1 = { error: e_1_1 }; }
82
89
  finally {
83
90
  try {
84
- if (consumer_1_1 && !consumer_1_1.done && (_a = consumer_1.return)) yield _a.call(consumer_1);
91
+ if (!_e && !_a && (_b = consumer_1.return)) yield _b.call(consumer_1);
85
92
  }
86
93
  finally { if (e_1) throw e_1.error; }
87
94
  }
@@ -82,9 +82,9 @@ function getInput(filename) {
82
82
  });
83
83
  }
84
84
  function produce(topic, opts, { parent }) {
85
- var e_1, _a;
85
+ var _a, e_1, _b, _c;
86
86
  return __awaiter(this, void 0, void 0, function* () {
87
- const _b = Object.assign(Object.assign({}, parent.opts()), opts), { format, header, input: filename, delay, brokers, logLevel, ssl } = _b, rest = __rest(_b, ["format", "header", "input", "delay", "brokers", "logLevel", "ssl"]);
87
+ const _d = Object.assign(Object.assign({}, parent.opts()), opts), { format, header, input: filename, delay, brokers, logLevel, ssl } = _d, rest = __rest(_d, ["format", "header", "input", "delay", "brokers", "logLevel", "ssl"]);
88
88
  const sasl = (0, kafka_1.getSASL)(rest);
89
89
  const client = (0, kafka_1.createClient)(brokers, ssl, sasl, logLevel);
90
90
  const producer = yield (0, kafka_1.createProducer)(client, topic);
@@ -101,18 +101,25 @@ function produce(topic, opts, { parent }) {
101
101
  const formatter = (0, formatters_1.getFormatter)(format);
102
102
  const input = yield getInput(filename);
103
103
  try {
104
- for (var input_1 = __asyncValues(input), input_1_1; input_1_1 = yield input_1.next(), !input_1_1.done;) {
105
- let { key, value, headers } = input_1_1.value;
106
- const encodedValue = yield formatter.encode(value);
107
- const message = { key, value: encodedValue, headers: Object.assign(Object.assign({}, staticHeaders), headers) };
108
- producer.push(message);
109
- yield new Promise(resolve => setTimeout(resolve, delay));
104
+ for (var _e = true, input_1 = __asyncValues(input), input_1_1; input_1_1 = yield input_1.next(), _a = input_1_1.done, !_a;) {
105
+ _c = input_1_1.value;
106
+ _e = false;
107
+ try {
108
+ let { key, value, headers } = _c;
109
+ const encodedValue = yield formatter.encode(value);
110
+ const message = { key, value: encodedValue, headers: Object.assign(Object.assign({}, staticHeaders), headers) };
111
+ producer.push(message);
112
+ yield new Promise(resolve => setTimeout(resolve, delay));
113
+ }
114
+ finally {
115
+ _e = true;
116
+ }
110
117
  }
111
118
  }
112
119
  catch (e_1_1) { e_1 = { error: e_1_1 }; }
113
120
  finally {
114
121
  try {
115
- if (input_1_1 && !input_1_1.done && (_a = input_1.return)) yield _a.call(input_1);
122
+ if (!_e && !_a && (_b = input_1.return)) yield _b.call(input_1);
116
123
  }
117
124
  finally { if (e_1) throw e_1.error; }
118
125
  }
@@ -142,20 +142,27 @@ function createProducer(client, topic) {
142
142
  producer.disconnect();
143
143
  });
144
144
  (() => __awaiter(this, void 0, void 0, function* () {
145
- var e_1, _a;
145
+ var _a, e_1, _b, _c;
146
146
  try {
147
- for (var pool_2 = __asyncValues(pool), pool_2_1; pool_2_1 = yield pool_2.next(), !pool_2_1.done;) {
148
- let message = pool_2_1.value;
149
- yield producer.send({
150
- topic,
151
- messages: [message],
152
- });
147
+ for (var _d = true, pool_2 = __asyncValues(pool), pool_2_1; pool_2_1 = yield pool_2.next(), _a = pool_2_1.done, !_a;) {
148
+ _c = pool_2_1.value;
149
+ _d = false;
150
+ try {
151
+ let message = _c;
152
+ yield producer.send({
153
+ topic,
154
+ messages: [message],
155
+ });
156
+ }
157
+ finally {
158
+ _d = true;
159
+ }
153
160
  }
154
161
  }
155
162
  catch (e_1_1) { e_1 = { error: e_1_1 }; }
156
163
  finally {
157
164
  try {
158
- if (pool_2_1 && !pool_2_1.done && (_a = pool_2.return)) yield _a.call(pool_2);
165
+ if (!_d && !_a && (_b = pool_2.return)) yield _b.call(pool_2);
159
166
  }
160
167
  finally { if (e_1) throw e_1.error; }
161
168
  }
@@ -68,19 +68,26 @@ class Pool {
68
68
  return this;
69
69
  }
70
70
  toArray() {
71
- var e_1, _a;
71
+ var _a, e_1, _b, _c;
72
72
  return __awaiter(this, void 0, void 0, function* () {
73
73
  const result = [];
74
74
  try {
75
- for (var _b = __asyncValues(this), _c; _c = yield _b.next(), !_c.done;) {
76
- let item = _c.value;
77
- result.push(item);
75
+ for (var _d = true, _e = __asyncValues(this), _f; _f = yield _e.next(), _a = _f.done, !_a;) {
76
+ _c = _f.value;
77
+ _d = false;
78
+ try {
79
+ let item = _c;
80
+ result.push(item);
81
+ }
82
+ finally {
83
+ _d = true;
84
+ }
78
85
  }
79
86
  }
80
87
  catch (e_1_1) { e_1 = { error: e_1_1 }; }
81
88
  finally {
82
89
  try {
83
- if (_c && !_c.done && (_a = _b.return)) yield _a.call(_b);
90
+ if (!_d && !_a && (_b = _e.return)) yield _b.call(_e);
84
91
  }
85
92
  finally { if (e_1) throw e_1.error; }
86
93
  }
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "kafka-console",
3
- "version": "1.2.58",
3
+ "version": "1.2.59",
4
4
  "description": "Kafka CLI tool",
5
5
  "main": "index.js",
6
6
  "bin": {