@simfinity/constellation-client 1.0.7 → 1.0.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -37,20 +37,20 @@ var WebClient = class {
37
37
  * be closed to release the context on server side. See endSession().
38
38
  *
39
39
  * @remarks
40
- * A session MUST exist to connect the stream.
40
+ * A session MUST exist first in order to connect the stream next.
41
41
  *
42
- * @param instructions to the model, added to its context. This is the "system" input instructions.
43
- * @param audio whether to allow audio streaming or text-only
42
+ * @param audioEnabled whether this session can receive & produce audio as well as text
43
+ * @param voiceName LLM specific voice name e.g. with OpenAI this could be 'alloy'
44
44
  *
45
45
  * @exception
46
46
  * This method throws new Error(...) if unable to execute successfully for any reason.
47
- *
48
- * @example
49
- * ```TypeScript
50
- * await startSession("You are useful assistant", true)
51
- * ```
52
47
  */
53
- async startSession(instructions = "", audio = false) {
48
+ async startSession(audioEnabled, voiceName) {
49
+ const prepareBody = {
50
+ llmProvider: this.config.llm,
51
+ audioEnabled,
52
+ voiceName
53
+ };
54
54
  const response = await fetch(`${this.config.sessionEndpoint}/prepare_session`, {
55
55
  method: "POST",
56
56
  headers: {
@@ -58,11 +58,7 @@ var WebClient = class {
58
58
  "Content-Type": "application/json",
59
59
  "Accept": "application/json"
60
60
  },
61
- body: JSON.stringify({
62
- llmProvider: this.config.llm,
63
- systemPrompt: instructions,
64
- audio
65
- })
61
+ body: JSON.stringify(prepareBody)
66
62
  });
67
63
  if (!response.ok) {
68
64
  throw new Error(`Could not create a new chat session
@@ -116,7 +112,7 @@ var WebClient = class {
116
112
  * This method not only opens a websocket connection with the server but also initiates a
117
113
  * handshake, where the server explicitly acknowledges and accepts the client connection.
118
114
  *
119
- * @param audio for a session that created with audio capabilities, allows this streaming to include audio
115
+ * @param audio for an audio-enabled session, request this streaming connection to include audio events
120
116
  * @param handlers callback functions to handle every possible communication events coming from the server
121
117
  *
122
118
  * @exception
@@ -125,7 +121,7 @@ var WebClient = class {
125
121
  * @example
126
122
  * ```TypeScript
127
123
  * // Open an audio connection: in this example we choose to handle only audio data, and ignore text.
128
- * await connect(true, {
124
+ * await joinSession(true, {
129
125
  * onStreamClosed: (reason: string) => { console.log("Stream connection lost"); },
130
126
  * onAudioResponseStart: () => { console.log("The model is talking"); },
131
127
  * onAudioResponseChunk: (audioChunk: string) => { audioPlayer.enqueue(audioChunk); },
@@ -187,7 +183,7 @@ var WebClient = class {
187
183
  this.ws = ws;
188
184
  }
189
185
  /**
190
- * With an opened streaming connection: send a "system" update of the session settings.
186
+ * Once a session is joined: send a "system" update of the session settings.
191
187
  * Allows to change some behavioral parameters like the temperature or system instructions.
192
188
  * This does not trigger a model response.
193
189
  *
@@ -195,16 +191,16 @@ var WebClient = class {
195
191
  * With openai for example, this triggers (pseudo-code):
196
192
  * webSocket.send({ type: "session.update", session: { ... }})
197
193
  *
198
- * @param settings complete definition of the settings. Same values must be re-provided for unchanged parameters.
194
+ * @param settings new system settings to apply. Omitted values will remain unchanged
199
195
  *
200
196
  * @exception
201
197
  * This method throws new Error(...) if unable to execute successfully for any reason.
202
198
  */
203
- async configureSession(settings) {
199
+ configureSession(settings) {
204
200
  this.send("session.configure", settings);
205
201
  }
206
202
  /**
207
- * With an opened streaming connection: send a text input message to the LLM. This will trigger a
203
+ * Once a session is joined: send a text input message to the LLM. This will trigger a
208
204
  * text response as well as an audio response if the session was opened with audio mode active.
209
205
  *
210
206
  * @remarks
@@ -221,7 +217,7 @@ var WebClient = class {
221
217
  this.send("text.input.message", { text });
222
218
  }
223
219
  /**
224
- * With an opened streaming connection: send a chunk of raw audio data to the LLM.
220
+ * Once a session is joined: send a chunk of raw audio data to the LLM.
225
221
  * Audio data chunks do not systematically & immediately trigger a model response:
226
222
  * They get accumulated by the model to form a single input message, until:
227
223
  * - commitAudioChunksSent is called, which flushes the accumulated audio and triggers a response
@@ -240,8 +236,8 @@ var WebClient = class {
240
236
  this.send("audio.input.append", { audioData: chunk });
241
237
  }
242
238
  /**
243
- * With an opened streaming connection: triggers the processing of the accumulated audio data since
244
- * the last model response. This effectively flushes the audio buffer and triggers a new model response.
239
+ * Once a session is joined: triggers the processing of the accumulated audio data since
240
+ * the last model response. This effectively flushes the audio buffer and triggers a model response.
245
241
  *
246
242
  * @remarks
247
243
  * With openai for example, this triggers (pseudo-code):
@@ -279,14 +275,22 @@ var WebClient = class {
279
275
  ws.onopen = () => {
280
276
  const eventSubs = audio ? [0 /* Text */, 1 /* Audio */] : [0 /* Text */];
281
277
  ws.send(JSON.stringify({
282
- type: "connection.request",
283
- data: { subscription: eventSubs }
278
+ type: "connection.initiate",
279
+ data: {
280
+ subscription: eventSubs,
281
+ settings: {
282
+ audio: true,
283
+ voice: "alloy",
284
+ temperature: 0.8,
285
+ instructions: ""
286
+ }
287
+ }
284
288
  }));
285
289
  };
286
290
  ws.onmessage = (event) => {
287
291
  try {
288
292
  const data = JSON.parse(event.data);
289
- if (data.type === "connection.accepted")
293
+ if (data.type === "connection.initiated")
290
294
  resolve(true);
291
295
  else
292
296
  reject(new Error(`Received unexpected event: ${data.type}`));
package/dist/index.d.cts CHANGED
@@ -31,17 +31,13 @@ interface WebClientConfig {
31
31
  }
32
32
  /**
33
33
  * System settings influencing the model behavior:
34
- * @audio: to activate voice conversation
35
- * @voice: depending on the LLM solution, this is the voice name to be used in audio
36
34
  * @temperature: LLM creativity factor in 0-1 range
37
35
  * @instructions: system instructions giving context, rules and directions to guide the LLM behavior
38
36
  * @maxResponseToken: 1-4096 value, maximum number of token used for a single response. Undefined means unlimited.
39
37
  */
40
- interface SessionSettings {
41
- audio: boolean;
42
- voice: string;
43
- temperature: number;
44
- instructions: string;
38
+ interface SessionConfig {
39
+ temperature?: number;
40
+ instructions?: string;
45
41
  maxResponseToken?: number;
46
42
  }
47
43
  /**
@@ -66,7 +62,7 @@ interface SessionSettings {
66
62
  */
67
63
  interface EventHandlers {
68
64
  onStreamClosed: (reason: string) => void;
69
- onSessionConfigured?: (settings: SessionSettings) => void;
65
+ onSessionConfigured?: (settings: SessionConfig) => void;
70
66
  onAudioResponseStart?: () => void;
71
67
  onAudioResponseChunk?: (audioChunk: string) => void;
72
68
  onAudioResponseEnd?: () => void;
@@ -93,20 +89,15 @@ declare class WebClient {
93
89
  * be closed to release the context on server side. See endSession().
94
90
  *
95
91
  * @remarks
96
- * A session MUST exist to connect the stream.
92
+ * A session MUST exist first in order to connect the stream next.
97
93
  *
98
- * @param instructions to the model, added to its context. This is the "system" input instructions.
99
- * @param audio whether to allow audio streaming or text-only
94
+ * @param audioEnabled whether this session can receive & produce audio as well as text
95
+ * @param voiceName LLM specific voice name e.g. with OpenAI this could be 'alloy'
100
96
  *
101
97
  * @exception
102
98
  * This method throws new Error(...) if unable to execute successfully for any reason.
103
- *
104
- * @example
105
- * ```TypeScript
106
- * await startSession("You are useful assistant", true)
107
- * ```
108
99
  */
109
- startSession(instructions?: string, audio?: boolean): Promise<void>;
100
+ startSession(audioEnabled: boolean, voiceName?: string): Promise<void>;
110
101
  /**
111
102
  * Close an opened, persistent chat room, effectively killing the streaming as well if still opened.
112
103
  * If there is no active session, this method does nothing.
@@ -128,7 +119,7 @@ declare class WebClient {
128
119
  * This method not only opens a websocket connection with the server but also initiates a
129
120
  * handshake, where the server explicitly acknowledges and accepts the client connection.
130
121
  *
131
- * @param audio for a session that created with audio capabilities, allows this streaming to include audio
122
+ * @param audio for an audio-enabled session, request this streaming connection to include audio events
132
123
  * @param handlers callback functions to handle every possible communication events coming from the server
133
124
  *
134
125
  * @exception
@@ -137,7 +128,7 @@ declare class WebClient {
137
128
  * @example
138
129
  * ```TypeScript
139
130
  * // Open an audio connection: in this example we choose to handle only audio data, and ignore text.
140
- * await connect(true, {
131
+ * await joinSession(true, {
141
132
  * onStreamClosed: (reason: string) => { console.log("Stream connection lost"); },
142
133
  * onAudioResponseStart: () => { console.log("The model is talking"); },
143
134
  * onAudioResponseChunk: (audioChunk: string) => { audioPlayer.enqueue(audioChunk); },
@@ -147,7 +138,7 @@ declare class WebClient {
147
138
  */
148
139
  joinSession(audio: boolean | undefined, handlers: EventHandlers): Promise<void>;
149
140
  /**
150
- * With an opened streaming connection: send a "system" update of the session settings.
141
+ * Once a session is joined: send a "system" update of the session settings.
151
142
  * Allows to change some behavioral parameters like the temperature or system instructions.
152
143
  * This does not trigger a model response.
153
144
  *
@@ -155,14 +146,14 @@ declare class WebClient {
155
146
  * With openai for example, this triggers (pseudo-code):
156
147
  * webSocket.send({ type: "session.update", session: { ... }})
157
148
  *
158
- * @param settings complete definition of the settings. Same values must be re-provided for unchanged parameters.
149
+ * @param settings new system settings to apply. Omitted values will remain unchanged
159
150
  *
160
151
  * @exception
161
152
  * This method throws new Error(...) if unable to execute successfully for any reason.
162
153
  */
163
- configureSession(settings: SessionSettings): Promise<void>;
154
+ configureSession(settings: SessionConfig): void;
164
155
  /**
165
- * With an opened streaming connection: send a text input message to the LLM. This will trigger a
156
+ * Once a session is joined: send a text input message to the LLM. This will trigger a
166
157
  * text response as well as an audio response if the session was opened with audio mode active.
167
158
  *
168
159
  * @remarks
@@ -177,7 +168,7 @@ declare class WebClient {
177
168
  */
178
169
  sendText(text: string): void;
179
170
  /**
180
- * With an opened streaming connection: send a chunk of raw audio data to the LLM.
171
+ * Once a session is joined: send a chunk of raw audio data to the LLM.
181
172
  * Audio data chunks do not systematically & immediately trigger a model response:
182
173
  * They get accumulated by the model to form a single input message, until:
183
174
  * - commitAudioChunksSent is called, which flushes the accumulated audio and triggers a response
@@ -194,8 +185,8 @@ declare class WebClient {
194
185
  */
195
186
  sendAudioChunk(chunk: string): void;
196
187
  /**
197
- * With an opened streaming connection: triggers the processing of the accumulated audio data since
198
- * the last model response. This effectively flushes the audio buffer and triggers a new model response.
188
+ * Once a session is joined: triggers the processing of the accumulated audio data since
189
+ * the last model response. This effectively flushes the audio buffer and triggers a model response.
199
190
  *
200
191
  * @remarks
201
192
  * With openai for example, this triggers (pseudo-code):
@@ -215,4 +206,4 @@ declare class WebClient {
215
206
  private send;
216
207
  }
217
208
 
218
- export { type EventHandlers, type LlmType, type SessionSettings, WebClient, type WebClientConfig };
209
+ export { type EventHandlers, type LlmType, type SessionConfig, WebClient, type WebClientConfig };
package/dist/index.d.ts CHANGED
@@ -31,17 +31,13 @@ interface WebClientConfig {
31
31
  }
32
32
  /**
33
33
  * System settings influencing the model behavior:
34
- * @audio: to activate voice conversation
35
- * @voice: depending on the LLM solution, this is the voice name to be used in audio
36
34
  * @temperature: LLM creativity factor in 0-1 range
37
35
  * @instructions: system instructions giving context, rules and directions to guide the LLM behavior
38
36
  * @maxResponseToken: 1-4096 value, maximum number of token used for a single response. Undefined means unlimited.
39
37
  */
40
- interface SessionSettings {
41
- audio: boolean;
42
- voice: string;
43
- temperature: number;
44
- instructions: string;
38
+ interface SessionConfig {
39
+ temperature?: number;
40
+ instructions?: string;
45
41
  maxResponseToken?: number;
46
42
  }
47
43
  /**
@@ -66,7 +62,7 @@ interface SessionSettings {
66
62
  */
67
63
  interface EventHandlers {
68
64
  onStreamClosed: (reason: string) => void;
69
- onSessionConfigured?: (settings: SessionSettings) => void;
65
+ onSessionConfigured?: (settings: SessionConfig) => void;
70
66
  onAudioResponseStart?: () => void;
71
67
  onAudioResponseChunk?: (audioChunk: string) => void;
72
68
  onAudioResponseEnd?: () => void;
@@ -93,20 +89,15 @@ declare class WebClient {
93
89
  * be closed to release the context on server side. See endSession().
94
90
  *
95
91
  * @remarks
96
- * A session MUST exist to connect the stream.
92
+ * A session MUST exist first in order to connect the stream next.
97
93
  *
98
- * @param instructions to the model, added to its context. This is the "system" input instructions.
99
- * @param audio whether to allow audio streaming or text-only
94
+ * @param audioEnabled whether this session can receive & produce audio as well as text
95
+ * @param voiceName LLM specific voice name e.g. with OpenAI this could be 'alloy'
100
96
  *
101
97
  * @exception
102
98
  * This method throws new Error(...) if unable to execute successfully for any reason.
103
- *
104
- * @example
105
- * ```TypeScript
106
- * await startSession("You are useful assistant", true)
107
- * ```
108
99
  */
109
- startSession(instructions?: string, audio?: boolean): Promise<void>;
100
+ startSession(audioEnabled: boolean, voiceName?: string): Promise<void>;
110
101
  /**
111
102
  * Close an opened, persistent chat room, effectively killing the streaming as well if still opened.
112
103
  * If there is no active session, this method does nothing.
@@ -128,7 +119,7 @@ declare class WebClient {
128
119
  * This method not only opens a websocket connection with the server but also initiates a
129
120
  * handshake, where the server explicitly acknowledges and accepts the client connection.
130
121
  *
131
- * @param audio for a session that created with audio capabilities, allows this streaming to include audio
122
+ * @param audio for an audio-enabled session, request this streaming connection to include audio events
132
123
  * @param handlers callback functions to handle every possible communication events coming from the server
133
124
  *
134
125
  * @exception
@@ -137,7 +128,7 @@ declare class WebClient {
137
128
  * @example
138
129
  * ```TypeScript
139
130
  * // Open an audio connection: in this example we choose to handle only audio data, and ignore text.
140
- * await connect(true, {
131
+ * await joinSession(true, {
141
132
  * onStreamClosed: (reason: string) => { console.log("Stream connection lost"); },
142
133
  * onAudioResponseStart: () => { console.log("The model is talking"); },
143
134
  * onAudioResponseChunk: (audioChunk: string) => { audioPlayer.enqueue(audioChunk); },
@@ -147,7 +138,7 @@ declare class WebClient {
147
138
  */
148
139
  joinSession(audio: boolean | undefined, handlers: EventHandlers): Promise<void>;
149
140
  /**
150
- * With an opened streaming connection: send a "system" update of the session settings.
141
+ * Once a session is joined: send a "system" update of the session settings.
151
142
  * Allows to change some behavioral parameters like the temperature or system instructions.
152
143
  * This does not trigger a model response.
153
144
  *
@@ -155,14 +146,14 @@ declare class WebClient {
155
146
  * With openai for example, this triggers (pseudo-code):
156
147
  * webSocket.send({ type: "session.update", session: { ... }})
157
148
  *
158
- * @param settings complete definition of the settings. Same values must be re-provided for unchanged parameters.
149
+ * @param settings new system settings to apply. Omitted values will remain unchanged
159
150
  *
160
151
  * @exception
161
152
  * This method throws new Error(...) if unable to execute successfully for any reason.
162
153
  */
163
- configureSession(settings: SessionSettings): Promise<void>;
154
+ configureSession(settings: SessionConfig): void;
164
155
  /**
165
- * With an opened streaming connection: send a text input message to the LLM. This will trigger a
156
+ * Once a session is joined: send a text input message to the LLM. This will trigger a
166
157
  * text response as well as an audio response if the session was opened with audio mode active.
167
158
  *
168
159
  * @remarks
@@ -177,7 +168,7 @@ declare class WebClient {
177
168
  */
178
169
  sendText(text: string): void;
179
170
  /**
180
- * With an opened streaming connection: send a chunk of raw audio data to the LLM.
171
+ * Once a session is joined: send a chunk of raw audio data to the LLM.
181
172
  * Audio data chunks do not systematically & immediately trigger a model response:
182
173
  * They get accumulated by the model to form a single input message, until:
183
174
  * - commitAudioChunksSent is called, which flushes the accumulated audio and triggers a response
@@ -194,8 +185,8 @@ declare class WebClient {
194
185
  */
195
186
  sendAudioChunk(chunk: string): void;
196
187
  /**
197
- * With an opened streaming connection: triggers the processing of the accumulated audio data since
198
- * the last model response. This effectively flushes the audio buffer and triggers a new model response.
188
+ * Once a session is joined: triggers the processing of the accumulated audio data since
189
+ * the last model response. This effectively flushes the audio buffer and triggers a model response.
199
190
  *
200
191
  * @remarks
201
192
  * With openai for example, this triggers (pseudo-code):
@@ -215,4 +206,4 @@ declare class WebClient {
215
206
  private send;
216
207
  }
217
208
 
218
- export { type EventHandlers, type LlmType, type SessionSettings, WebClient, type WebClientConfig };
209
+ export { type EventHandlers, type LlmType, type SessionConfig, WebClient, type WebClientConfig };
package/dist/index.js CHANGED
@@ -11,20 +11,20 @@ var WebClient = class {
11
11
  * be closed to release the context on server side. See endSession().
12
12
  *
13
13
  * @remarks
14
- * A session MUST exist to connect the stream.
14
+ * A session MUST exist first in order to connect the stream next.
15
15
  *
16
- * @param instructions to the model, added to its context. This is the "system" input instructions.
17
- * @param audio whether to allow audio streaming or text-only
16
+ * @param audioEnabled whether this session can receive & produce audio as well as text
17
+ * @param voiceName LLM specific voice name e.g. with OpenAI this could be 'alloy'
18
18
  *
19
19
  * @exception
20
20
  * This method throws new Error(...) if unable to execute successfully for any reason.
21
- *
22
- * @example
23
- * ```TypeScript
24
- * await startSession("You are useful assistant", true)
25
- * ```
26
21
  */
27
- async startSession(instructions = "", audio = false) {
22
+ async startSession(audioEnabled, voiceName) {
23
+ const prepareBody = {
24
+ llmProvider: this.config.llm,
25
+ audioEnabled,
26
+ voiceName
27
+ };
28
28
  const response = await fetch(`${this.config.sessionEndpoint}/prepare_session`, {
29
29
  method: "POST",
30
30
  headers: {
@@ -32,11 +32,7 @@ var WebClient = class {
32
32
  "Content-Type": "application/json",
33
33
  "Accept": "application/json"
34
34
  },
35
- body: JSON.stringify({
36
- llmProvider: this.config.llm,
37
- systemPrompt: instructions,
38
- audio
39
- })
35
+ body: JSON.stringify(prepareBody)
40
36
  });
41
37
  if (!response.ok) {
42
38
  throw new Error(`Could not create a new chat session
@@ -90,7 +86,7 @@ var WebClient = class {
90
86
  * This method not only opens a websocket connection with the server but also initiates a
91
87
  * handshake, where the server explicitly acknowledges and accepts the client connection.
92
88
  *
93
- * @param audio for a session that created with audio capabilities, allows this streaming to include audio
89
+ * @param audio for an audio-enabled session, request this streaming connection to include audio events
94
90
  * @param handlers callback functions to handle every possible communication events coming from the server
95
91
  *
96
92
  * @exception
@@ -99,7 +95,7 @@ var WebClient = class {
99
95
  * @example
100
96
  * ```TypeScript
101
97
  * // Open an audio connection: in this example we choose to handle only audio data, and ignore text.
102
- * await connect(true, {
98
+ * await joinSession(true, {
103
99
  * onStreamClosed: (reason: string) => { console.log("Stream connection lost"); },
104
100
  * onAudioResponseStart: () => { console.log("The model is talking"); },
105
101
  * onAudioResponseChunk: (audioChunk: string) => { audioPlayer.enqueue(audioChunk); },
@@ -161,7 +157,7 @@ var WebClient = class {
161
157
  this.ws = ws;
162
158
  }
163
159
  /**
164
- * With an opened streaming connection: send a "system" update of the session settings.
160
+ * Once a session is joined: send a "system" update of the session settings.
165
161
  * Allows to change some behavioral parameters like the temperature or system instructions.
166
162
  * This does not trigger a model response.
167
163
  *
@@ -169,16 +165,16 @@ var WebClient = class {
169
165
  * With openai for example, this triggers (pseudo-code):
170
166
  * webSocket.send({ type: "session.update", session: { ... }})
171
167
  *
172
- * @param settings complete definition of the settings. Same values must be re-provided for unchanged parameters.
168
+ * @param settings new system settings to apply. Omitted values will remain unchanged
173
169
  *
174
170
  * @exception
175
171
  * This method throws new Error(...) if unable to execute successfully for any reason.
176
172
  */
177
- async configureSession(settings) {
173
+ configureSession(settings) {
178
174
  this.send("session.configure", settings);
179
175
  }
180
176
  /**
181
- * With an opened streaming connection: send a text input message to the LLM. This will trigger a
177
+ * Once a session is joined: send a text input message to the LLM. This will trigger a
182
178
  * text response as well as an audio response if the session was opened with audio mode active.
183
179
  *
184
180
  * @remarks
@@ -195,7 +191,7 @@ var WebClient = class {
195
191
  this.send("text.input.message", { text });
196
192
  }
197
193
  /**
198
- * With an opened streaming connection: send a chunk of raw audio data to the LLM.
194
+ * Once a session is joined: send a chunk of raw audio data to the LLM.
199
195
  * Audio data chunks do not systematically & immediately trigger a model response:
200
196
  * They get accumulated by the model to form a single input message, until:
201
197
  * - commitAudioChunksSent is called, which flushes the accumulated audio and triggers a response
@@ -214,8 +210,8 @@ var WebClient = class {
214
210
  this.send("audio.input.append", { audioData: chunk });
215
211
  }
216
212
  /**
217
- * With an opened streaming connection: triggers the processing of the accumulated audio data since
218
- * the last model response. This effectively flushes the audio buffer and triggers a new model response.
213
+ * Once a session is joined: triggers the processing of the accumulated audio data since
214
+ * the last model response. This effectively flushes the audio buffer and triggers a model response.
219
215
  *
220
216
  * @remarks
221
217
  * With openai for example, this triggers (pseudo-code):
@@ -253,14 +249,22 @@ var WebClient = class {
253
249
  ws.onopen = () => {
254
250
  const eventSubs = audio ? [0 /* Text */, 1 /* Audio */] : [0 /* Text */];
255
251
  ws.send(JSON.stringify({
256
- type: "connection.request",
257
- data: { subscription: eventSubs }
252
+ type: "connection.initiate",
253
+ data: {
254
+ subscription: eventSubs,
255
+ settings: {
256
+ audio: true,
257
+ voice: "alloy",
258
+ temperature: 0.8,
259
+ instructions: ""
260
+ }
261
+ }
258
262
  }));
259
263
  };
260
264
  ws.onmessage = (event) => {
261
265
  try {
262
266
  const data = JSON.parse(event.data);
263
- if (data.type === "connection.accepted")
267
+ if (data.type === "connection.initiated")
264
268
  resolve(true);
265
269
  else
266
270
  reject(new Error(`Received unexpected event: ${data.type}`));
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@simfinity/constellation-client",
3
- "version": "1.0.7",
3
+ "version": "1.0.8",
4
4
  "type": "module",
5
5
  "exports": {
6
6
  ".": {