@smartytalent/openai-tools 0.1.33-dev.53 → 0.1.33-dev.55

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (2) hide show
  1. package/dist/tools.json +372 -13
  2. package/package.json +1 -1
package/dist/tools.json CHANGED
@@ -14060,6 +14060,189 @@
14060
14060
  "type": "object"
14061
14061
  }
14062
14062
  },
14063
+ "recordingStatus": {
14064
+ "type": "string",
14065
+ "enum": [
14066
+ "recording",
14067
+ "success",
14068
+ "failed"
14069
+ ],
14070
+ "description": "Chime SDK Media Capture Pipeline lifecycle. Set by the Boe bot when it starts/stops the capture."
14071
+ },
14072
+ "mediaCapturePipelineArn": {
14073
+ "type": "string",
14074
+ "description": "Chime SDK Media Capture Pipeline ARN. Internal; FE shouldn't need this directly."
14075
+ },
14076
+ "recordingS3Prefix": {
14077
+ "type": "string",
14078
+ "description": "S3 key prefix where capture pipeline wrote composited-video chunks. Inside the meeting service bucket."
14079
+ },
14080
+ "recordingStartedAt": {
14081
+ "type": "string",
14082
+ "format": "date-time"
14083
+ },
14084
+ "recordingEndedAt": {
14085
+ "type": "string",
14086
+ "format": "date-time"
14087
+ },
14088
+ "recordingError": {
14089
+ "type": "string",
14090
+ "description": "Set when recordingStatus=failed. Includes the AWS error name + message."
14091
+ },
14092
+ "concatenationStatus": {
14093
+ "type": "string",
14094
+ "enum": [
14095
+ "running",
14096
+ "success",
14097
+ "failed"
14098
+ ],
14099
+ "description": "Chime SDK Media Concatenation Pipeline lifecycle. Concat merges the ~5s capture chunks into one composited-video.mp4."
14100
+ },
14101
+ "concatenationPipelineArn": {
14102
+ "type": "string",
14103
+ "description": "Concat pipeline ARN. Last path component is the {concatId} used as the merged MP4 filename."
14104
+ },
14105
+ "concatenationS3Prefix": {
14106
+ "type": "string",
14107
+ "description": "S3 key prefix where the concat output lands. Usually {shardId}/{meetingId}/merged/."
14108
+ },
14109
+ "concatenationStartedAt": {
14110
+ "type": "string",
14111
+ "format": "date-time"
14112
+ },
14113
+ "concatenationEndedAt": {
14114
+ "type": "string",
14115
+ "format": "date-time"
14116
+ },
14117
+ "concatenationError": {
14118
+ "type": "string"
14119
+ },
14120
+ "aiProcessingStatus": {
14121
+ "type": "string",
14122
+ "enum": [
14123
+ "running",
14124
+ "success",
14125
+ "failed"
14126
+ ],
14127
+ "description": "Phase 3 Step Function (ffmpeg + gpt-audio + translate + finalize) lifecycle. Set by the finalize Lambda."
14128
+ },
14129
+ "aiProcessingStartedAt": {
14130
+ "type": "string",
14131
+ "format": "date-time"
14132
+ },
14133
+ "aiProcessingEndedAt": {
14134
+ "type": "string",
14135
+ "format": "date-time"
14136
+ },
14137
+ "aiProcessingError": {
14138
+ "type": "string"
14139
+ },
14140
+ "transcriptSegments": {
14141
+ "type": "array",
14142
+ "description": "Time-coded transcript turns. Same shape as call transcriptSegments + adds attendeeId so FE can colour-code participants.",
14143
+ "items": {
14144
+ "title": "MeetingTranscriptSegmentSchema",
14145
+ "type": "object",
14146
+ "description": "Single speaker turn in the meeting transcript. Mirrors CallTranscriptSegmentSchema + attendeeId.",
14147
+ "properties": {
14148
+ "startMs": {
14149
+ "type": "integer",
14150
+ "description": "Offset from start of recording, milliseconds."
14151
+ },
14152
+ "endMs": {
14153
+ "type": "integer"
14154
+ },
14155
+ "speakerType": {
14156
+ "type": "string",
14157
+ "enum": [
14158
+ "ai",
14159
+ "human"
14160
+ ]
14161
+ },
14162
+ "speakerName": {
14163
+ "type": "string",
14164
+ "description": "Display name resolved from the meeting row (aiAttendees + participants), with a Speaker {id-suffix} fallback for unrecognised attendees."
14165
+ },
14166
+ "attendeeId": {
14167
+ "type": "string",
14168
+ "description": "chimeAttendeeId so FE can map back to the attendee record (avatar / colour / etc)."
14169
+ },
14170
+ "text": {
14171
+ "type": "string",
14172
+ "description": "Verbatim spoken text in the source language. NOT translated; use transcript[locale] for translated joined-text views."
14173
+ }
14174
+ }
14175
+ }
14176
+ },
14177
+ "speakerStats": {
14178
+ "type": "array",
14179
+ "description": "Per-speaker talk percentage in basis points (10000 = 100%).",
14180
+ "items": {
14181
+ "title": "MeetingSpeakerStatSchema",
14182
+ "type": "object",
14183
+ "properties": {
14184
+ "attendeeId": {
14185
+ "type": "string"
14186
+ },
14187
+ "speakerType": {
14188
+ "type": "string",
14189
+ "enum": [
14190
+ "ai",
14191
+ "human"
14192
+ ]
14193
+ },
14194
+ "speakerName": {
14195
+ "type": "string"
14196
+ },
14197
+ "talkPercentageBp": {
14198
+ "type": "integer",
14199
+ "description": "Talk percentage in basis points (10000 = 100%)."
14200
+ }
14201
+ }
14202
+ }
14203
+ },
14204
+ "languageCode": {
14205
+ "type": "string",
14206
+ "description": "BCP-47 primary language of the meeting, auto-detected by gpt-audio."
14207
+ },
14208
+ "summary": {
14209
+ "type": "string",
14210
+ "description": "AI-generated 2-4 paragraph summary in the source languageCode. Use summaryByLocale for translated copies."
14211
+ },
14212
+ "summaryByLocale": {
14213
+ "type": "object",
14214
+ "description": "Translated summary per tenant locale. Shape: { 'pl-PL': '...', 'en-US': '...' }.",
14215
+ "additionalProperties": {
14216
+ "type": "string"
14217
+ }
14218
+ },
14219
+ "summaryGeneratedAt": {
14220
+ "type": "string",
14221
+ "format": "date-time"
14222
+ },
14223
+ "keyNotes": {
14224
+ "type": "array",
14225
+ "description": "3-5 prose bullets distilling the meeting's substance, source language. NOT translated.",
14226
+ "items": {
14227
+ "type": "string"
14228
+ }
14229
+ },
14230
+ "topicsDiscussed": {
14231
+ "type": "array",
14232
+ "description": "Stable facet keys from the meeting topic dictionary (FE renders human labels via shared dict).",
14233
+ "items": {
14234
+ "type": "string"
14235
+ }
14236
+ },
14237
+ "sentiment": {
14238
+ "type": "string",
14239
+ "enum": [
14240
+ "positive",
14241
+ "neutral",
14242
+ "negative",
14243
+ "mixed"
14244
+ ]
14245
+ },
14063
14246
  "status": {
14064
14247
  "type": "string",
14065
14248
  "enum": [
@@ -14365,6 +14548,189 @@
14365
14548
  "type": "object"
14366
14549
  }
14367
14550
  },
14551
+ "recordingStatus": {
14552
+ "type": "string",
14553
+ "enum": [
14554
+ "recording",
14555
+ "success",
14556
+ "failed"
14557
+ ],
14558
+ "description": "Chime SDK Media Capture Pipeline lifecycle. Set by the Boe bot when it starts/stops the capture."
14559
+ },
14560
+ "mediaCapturePipelineArn": {
14561
+ "type": "string",
14562
+ "description": "Chime SDK Media Capture Pipeline ARN. Internal; FE shouldn't need this directly."
14563
+ },
14564
+ "recordingS3Prefix": {
14565
+ "type": "string",
14566
+ "description": "S3 key prefix where capture pipeline wrote composited-video chunks. Inside the meeting service bucket."
14567
+ },
14568
+ "recordingStartedAt": {
14569
+ "type": "string",
14570
+ "format": "date-time"
14571
+ },
14572
+ "recordingEndedAt": {
14573
+ "type": "string",
14574
+ "format": "date-time"
14575
+ },
14576
+ "recordingError": {
14577
+ "type": "string",
14578
+ "description": "Set when recordingStatus=failed. Includes the AWS error name + message."
14579
+ },
14580
+ "concatenationStatus": {
14581
+ "type": "string",
14582
+ "enum": [
14583
+ "running",
14584
+ "success",
14585
+ "failed"
14586
+ ],
14587
+ "description": "Chime SDK Media Concatenation Pipeline lifecycle. Concat merges the ~5s capture chunks into one composited-video.mp4."
14588
+ },
14589
+ "concatenationPipelineArn": {
14590
+ "type": "string",
14591
+ "description": "Concat pipeline ARN. Last path component is the {concatId} used as the merged MP4 filename."
14592
+ },
14593
+ "concatenationS3Prefix": {
14594
+ "type": "string",
14595
+ "description": "S3 key prefix where the concat output lands. Usually {shardId}/{meetingId}/merged/."
14596
+ },
14597
+ "concatenationStartedAt": {
14598
+ "type": "string",
14599
+ "format": "date-time"
14600
+ },
14601
+ "concatenationEndedAt": {
14602
+ "type": "string",
14603
+ "format": "date-time"
14604
+ },
14605
+ "concatenationError": {
14606
+ "type": "string"
14607
+ },
14608
+ "aiProcessingStatus": {
14609
+ "type": "string",
14610
+ "enum": [
14611
+ "running",
14612
+ "success",
14613
+ "failed"
14614
+ ],
14615
+ "description": "Phase 3 Step Function (ffmpeg + gpt-audio + translate + finalize) lifecycle. Set by the finalize Lambda."
14616
+ },
14617
+ "aiProcessingStartedAt": {
14618
+ "type": "string",
14619
+ "format": "date-time"
14620
+ },
14621
+ "aiProcessingEndedAt": {
14622
+ "type": "string",
14623
+ "format": "date-time"
14624
+ },
14625
+ "aiProcessingError": {
14626
+ "type": "string"
14627
+ },
14628
+ "transcriptSegments": {
14629
+ "type": "array",
14630
+ "description": "Time-coded transcript turns. Same shape as call transcriptSegments + adds attendeeId so FE can colour-code participants.",
14631
+ "items": {
14632
+ "title": "MeetingTranscriptSegmentSchema",
14633
+ "type": "object",
14634
+ "description": "Single speaker turn in the meeting transcript. Mirrors CallTranscriptSegmentSchema + attendeeId.",
14635
+ "properties": {
14636
+ "startMs": {
14637
+ "type": "integer",
14638
+ "description": "Offset from start of recording, milliseconds."
14639
+ },
14640
+ "endMs": {
14641
+ "type": "integer"
14642
+ },
14643
+ "speakerType": {
14644
+ "type": "string",
14645
+ "enum": [
14646
+ "ai",
14647
+ "human"
14648
+ ]
14649
+ },
14650
+ "speakerName": {
14651
+ "type": "string",
14652
+ "description": "Display name resolved from the meeting row (aiAttendees + participants), with a Speaker {id-suffix} fallback for unrecognised attendees."
14653
+ },
14654
+ "attendeeId": {
14655
+ "type": "string",
14656
+ "description": "chimeAttendeeId so FE can map back to the attendee record (avatar / colour / etc)."
14657
+ },
14658
+ "text": {
14659
+ "type": "string",
14660
+ "description": "Verbatim spoken text in the source language. NOT translated; use transcript[locale] for translated joined-text views."
14661
+ }
14662
+ }
14663
+ }
14664
+ },
14665
+ "speakerStats": {
14666
+ "type": "array",
14667
+ "description": "Per-speaker talk percentage in basis points (10000 = 100%).",
14668
+ "items": {
14669
+ "title": "MeetingSpeakerStatSchema",
14670
+ "type": "object",
14671
+ "properties": {
14672
+ "attendeeId": {
14673
+ "type": "string"
14674
+ },
14675
+ "speakerType": {
14676
+ "type": "string",
14677
+ "enum": [
14678
+ "ai",
14679
+ "human"
14680
+ ]
14681
+ },
14682
+ "speakerName": {
14683
+ "type": "string"
14684
+ },
14685
+ "talkPercentageBp": {
14686
+ "type": "integer",
14687
+ "description": "Talk percentage in basis points (10000 = 100%)."
14688
+ }
14689
+ }
14690
+ }
14691
+ },
14692
+ "languageCode": {
14693
+ "type": "string",
14694
+ "description": "BCP-47 primary language of the meeting, auto-detected by gpt-audio."
14695
+ },
14696
+ "summary": {
14697
+ "type": "string",
14698
+ "description": "AI-generated 2-4 paragraph summary in the source languageCode. Use summaryByLocale for translated copies."
14699
+ },
14700
+ "summaryByLocale": {
14701
+ "type": "object",
14702
+ "description": "Translated summary per tenant locale. Shape: { 'pl-PL': '...', 'en-US': '...' }.",
14703
+ "additionalProperties": {
14704
+ "type": "string"
14705
+ }
14706
+ },
14707
+ "summaryGeneratedAt": {
14708
+ "type": "string",
14709
+ "format": "date-time"
14710
+ },
14711
+ "keyNotes": {
14712
+ "type": "array",
14713
+ "description": "3-5 prose bullets distilling the meeting's substance, source language. NOT translated.",
14714
+ "items": {
14715
+ "type": "string"
14716
+ }
14717
+ },
14718
+ "topicsDiscussed": {
14719
+ "type": "array",
14720
+ "description": "Stable facet keys from the meeting topic dictionary (FE renders human labels via shared dict).",
14721
+ "items": {
14722
+ "type": "string"
14723
+ }
14724
+ },
14725
+ "sentiment": {
14726
+ "type": "string",
14727
+ "enum": [
14728
+ "positive",
14729
+ "neutral",
14730
+ "negative",
14731
+ "mixed"
14732
+ ]
14733
+ },
14368
14734
  "status": {
14369
14735
  "type": "string",
14370
14736
  "enum": [
@@ -14594,24 +14960,17 @@
14594
14960
  "properties": {
14595
14961
  "mode": {
14596
14962
  "type": "string",
14597
- "description": "AI session mode: translator or assistant",
14963
+ "description": "Boe's presence in the meeting. See MeetingAiSessionAttributesSchema.mode.",
14598
14964
  "enum": [
14599
- "translator",
14600
- "assistant"
14965
+ "listen",
14966
+ "voice",
14967
+ "face"
14601
14968
  ],
14602
- "default": "translator"
14603
- },
14604
- "sourceLanguage": {
14605
- "type": "string",
14606
- "description": "Source language for translation mode (e.g. Polish, English)"
14607
- },
14608
- "targetLanguage": {
14609
- "type": "string",
14610
- "description": "Target language for translation mode (e.g. English, German)"
14969
+ "default": "voice"
14611
14970
  },
14612
14971
  "language": {
14613
14972
  "type": "string",
14614
- "description": "Language for assistant mode (e.g. English)"
14973
+ "description": "Reply language for Boe's assistant behaviour (e.g. English)"
14615
14974
  }
14616
14975
  }
14617
14976
  }
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@smartytalent/openai-tools",
3
- "version": "0.1.33-dev.53",
3
+ "version": "0.1.33-dev.55",
4
4
  "description": "OpenAI function/tool definitions for SmartyTalent API",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",