telnyx 6.47.0 → 6.48.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/CHANGELOG.md +8 -0
  2. package/package.json +1 -1
  3. package/resources/ai/ai.d.mts +2 -2
  4. package/resources/ai/ai.d.mts.map +1 -1
  5. package/resources/ai/ai.d.ts +2 -2
  6. package/resources/ai/ai.d.ts.map +1 -1
  7. package/resources/ai/ai.js.map +1 -1
  8. package/resources/ai/ai.mjs.map +1 -1
  9. package/resources/ai/assistants/assistants.d.mts +140 -322
  10. package/resources/ai/assistants/assistants.d.mts.map +1 -1
  11. package/resources/ai/assistants/assistants.d.ts +140 -322
  12. package/resources/ai/assistants/assistants.d.ts.map +1 -1
  13. package/resources/ai/assistants/assistants.js.map +1 -1
  14. package/resources/ai/assistants/assistants.mjs.map +1 -1
  15. package/resources/ai/assistants/index.d.mts +1 -1
  16. package/resources/ai/assistants/index.d.mts.map +1 -1
  17. package/resources/ai/assistants/index.d.ts +1 -1
  18. package/resources/ai/assistants/index.d.ts.map +1 -1
  19. package/resources/ai/assistants/index.js.map +1 -1
  20. package/resources/ai/assistants/index.mjs.map +1 -1
  21. package/resources/ai/assistants/versions.d.mts +6 -214
  22. package/resources/ai/assistants/versions.d.mts.map +1 -1
  23. package/resources/ai/assistants/versions.d.ts +6 -214
  24. package/resources/ai/assistants/versions.d.ts.map +1 -1
  25. package/resources/ai/index.d.mts +1 -1
  26. package/resources/ai/index.d.mts.map +1 -1
  27. package/resources/ai/index.d.ts +1 -1
  28. package/resources/ai/index.d.ts.map +1 -1
  29. package/resources/ai/index.js.map +1 -1
  30. package/resources/ai/index.mjs.map +1 -1
  31. package/resources/calls/actions.d.mts +52 -4
  32. package/resources/calls/actions.d.mts.map +1 -1
  33. package/resources/calls/actions.d.ts +52 -4
  34. package/resources/calls/actions.d.ts.map +1 -1
  35. package/resources/conferences/actions.d.mts +13 -1
  36. package/resources/conferences/actions.d.mts.map +1 -1
  37. package/resources/conferences/actions.d.ts +13 -1
  38. package/resources/conferences/actions.d.ts.map +1 -1
  39. package/src/resources/ai/ai.ts +12 -0
  40. package/src/resources/ai/assistants/assistants.ts +167 -375
  41. package/src/resources/ai/assistants/index.ts +6 -0
  42. package/src/resources/ai/assistants/versions.ts +6 -250
  43. package/src/resources/ai/index.ts +6 -0
  44. package/src/resources/calls/actions.ts +64 -4
  45. package/src/resources/conferences/actions.ts +16 -1
  46. package/src/version.ts +1 -1
  47. package/version.d.mts +1 -1
  48. package/version.d.ts +1 -1
  49. package/version.js +1 -1
  50. package/version.mjs +1 -1
@@ -598,6 +598,104 @@ export interface AudioVisualizerConfig {
598
598
  * messages.
599
599
  */
600
600
  export type EnabledFeatures = 'telephony' | 'messaging';
601
+ export interface ExternalLlm {
602
+ /**
603
+ * Base URL for the external LLM endpoint.
604
+ */
605
+ base_url: string;
606
+ /**
607
+ * Model identifier to use with the external LLM endpoint.
608
+ */
609
+ model: string;
610
+ /**
611
+ * Authentication method used when connecting to the external LLM endpoint.
612
+ */
613
+ authentication_method?: 'token' | 'certificate';
614
+ /**
615
+ * Integration secret identifier for the client certificate used with certificate
616
+ * authentication.
617
+ */
618
+ certificate_ref?: string;
619
+ /**
620
+ * When `true`, Telnyx forwards the assistant's dynamic variables to the external
621
+ * LLM endpoint as a top-level `extra_metadata` object on the chat completion
622
+ * request body. Defaults to `false`. Example payload sent to the external
623
+ * endpoint:
624
+ * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
625
+ * Distinct from OpenAI's native `metadata` field, which has its own size and type
626
+ * limits.
627
+ */
628
+ forward_metadata?: boolean;
629
+ /**
630
+ * Integration secret identifier for the external LLM API key.
631
+ */
632
+ llm_api_key_ref?: string;
633
+ /**
634
+ * URL used to retrieve an access token when certificate authentication is enabled.
635
+ */
636
+ token_retrieval_url?: string;
637
+ }
638
+ export interface ExternalLlmReq {
639
+ /**
640
+ * Base URL for the external LLM endpoint.
641
+ */
642
+ base_url: string;
643
+ /**
644
+ * Model identifier to use with the external LLM endpoint.
645
+ */
646
+ model: string;
647
+ /**
648
+ * Authentication method used when connecting to the external LLM endpoint.
649
+ */
650
+ authentication_method?: 'token' | 'certificate';
651
+ /**
652
+ * Integration secret identifier for the client certificate used with certificate
653
+ * authentication.
654
+ */
655
+ certificate_ref?: string;
656
+ /**
657
+ * When `true`, Telnyx forwards the assistant's dynamic variables to the external
658
+ * LLM endpoint as a top-level `extra_metadata` object on the chat completion
659
+ * request body. Defaults to `false`. Example payload sent to the external
660
+ * endpoint:
661
+ * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
662
+ * Distinct from OpenAI's native `metadata` field, which has its own size and type
663
+ * limits.
664
+ */
665
+ forward_metadata?: boolean;
666
+ /**
667
+ * Integration secret identifier for the external LLM API key.
668
+ */
669
+ llm_api_key_ref?: string;
670
+ /**
671
+ * URL used to retrieve an access token when certificate authentication is enabled.
672
+ */
673
+ token_retrieval_url?: string;
674
+ }
675
+ export interface FallbackConfig {
676
+ external_llm?: ExternalLlm;
677
+ /**
678
+ * Integration secret identifier for the fallback model API key.
679
+ */
680
+ llm_api_key_ref?: string;
681
+ /**
682
+ * Fallback Telnyx-hosted model to use when the primary LLM provider is
683
+ * unavailable.
684
+ */
685
+ model?: string;
686
+ }
687
+ export interface FallbackConfigReq {
688
+ external_llm?: ExternalLlmReq;
689
+ /**
690
+ * Integration secret identifier for the fallback model API key.
691
+ */
692
+ llm_api_key_ref?: string;
693
+ /**
694
+ * Fallback Telnyx-hosted model to use when the primary LLM provider is
695
+ * unavailable.
696
+ */
697
+ model?: string;
698
+ }
601
699
  export interface HangupTool {
602
700
  hangup: HangupToolParams;
603
701
  type: 'hangup';
@@ -661,8 +759,8 @@ export interface InferenceEmbedding {
661
759
  */
662
760
  dynamic_variables_webhook_url?: string;
663
761
  enabled_features?: Array<EnabledFeatures>;
664
- external_llm?: InferenceEmbedding.ExternalLlm;
665
- fallback_config?: InferenceEmbedding.FallbackConfig;
762
+ external_llm?: ExternalLlm;
763
+ fallback_config?: FallbackConfig;
666
764
  /**
667
765
  * Text that the assistant will use to start the conversation. This may be
668
766
  * templated with
@@ -714,7 +812,7 @@ export interface InferenceEmbedding {
714
812
  * Telephony-control tools (e.g. hangup, transfer) are unavailable
715
813
  * post-conversation. Beta feature.
716
814
  */
717
- post_conversation_settings?: InferenceEmbedding.PostConversationSettings;
815
+ post_conversation_settings?: PostConversationSettings;
718
816
  privacy_settings?: PrivacySettings;
719
817
  /**
720
818
  * IDs of missions related to this assistant.
@@ -753,94 +851,6 @@ export interface InferenceEmbedding {
753
851
  widget_settings?: WidgetSettings;
754
852
  }
755
853
  export declare namespace InferenceEmbedding {
756
- interface ExternalLlm {
757
- /**
758
- * Base URL for the external LLM endpoint.
759
- */
760
- base_url: string;
761
- /**
762
- * Model identifier to use with the external LLM endpoint.
763
- */
764
- model: string;
765
- /**
766
- * Authentication method used when connecting to the external LLM endpoint.
767
- */
768
- authentication_method?: 'token' | 'certificate';
769
- /**
770
- * Integration secret identifier for the client certificate used with certificate
771
- * authentication.
772
- */
773
- certificate_ref?: string;
774
- /**
775
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
776
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
777
- * request body. Defaults to `false`. Example payload sent to the external
778
- * endpoint:
779
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
780
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
781
- * limits.
782
- */
783
- forward_metadata?: boolean;
784
- /**
785
- * Integration secret identifier for the external LLM API key.
786
- */
787
- llm_api_key_ref?: string;
788
- /**
789
- * URL used to retrieve an access token when certificate authentication is enabled.
790
- */
791
- token_retrieval_url?: string;
792
- }
793
- interface FallbackConfig {
794
- external_llm?: FallbackConfig.ExternalLlm;
795
- /**
796
- * Integration secret identifier for the fallback model API key.
797
- */
798
- llm_api_key_ref?: string;
799
- /**
800
- * Fallback Telnyx-hosted model to use when the primary LLM provider is
801
- * unavailable.
802
- */
803
- model?: string;
804
- }
805
- namespace FallbackConfig {
806
- interface ExternalLlm {
807
- /**
808
- * Base URL for the external LLM endpoint.
809
- */
810
- base_url: string;
811
- /**
812
- * Model identifier to use with the external LLM endpoint.
813
- */
814
- model: string;
815
- /**
816
- * Authentication method used when connecting to the external LLM endpoint.
817
- */
818
- authentication_method?: 'token' | 'certificate';
819
- /**
820
- * Integration secret identifier for the client certificate used with certificate
821
- * authentication.
822
- */
823
- certificate_ref?: string;
824
- /**
825
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
826
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
827
- * request body. Defaults to `false`. Example payload sent to the external
828
- * endpoint:
829
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
830
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
831
- * limits.
832
- */
833
- forward_metadata?: boolean;
834
- /**
835
- * Integration secret identifier for the external LLM API key.
836
- */
837
- llm_api_key_ref?: string;
838
- /**
839
- * URL used to retrieve an access token when certificate authentication is enabled.
840
- */
841
- token_retrieval_url?: string;
842
- }
843
- }
844
854
  /**
845
855
  * Reference to a connected integration attached to an assistant. Discover
846
856
  * available integrations with `/ai/integrations` and connected integrations with
@@ -942,22 +952,6 @@ export declare namespace InferenceEmbedding {
942
952
  */
943
953
  allowed_tools?: Array<string>;
944
954
  }
945
- /**
946
- * Configuration for post-conversation processing. When enabled, the assistant
947
- * receives one additional LLM turn after the conversation ends, allowing it to
948
- * execute tool calls such as logging to a CRM or sending a summary. The assistant
949
- * can execute multiple parallel or sequential tools during this phase.
950
- * Telephony-control tools (e.g. hangup, transfer) are unavailable
951
- * post-conversation. Beta feature.
952
- */
953
- interface PostConversationSettings {
954
- /**
955
- * Whether post-conversation processing is enabled. When true, the assistant will
956
- * be invoked after the conversation ends to perform any final tool calls. Defaults
957
- * to false.
958
- */
959
- enabled?: boolean;
960
- }
961
955
  }
962
956
  export interface InferenceEmbeddingWebhookToolParams {
963
957
  type: 'webhook';
@@ -1171,6 +1165,38 @@ export interface ObservabilityReq {
1171
1165
  secret_key_ref?: string;
1172
1166
  status?: 'enabled' | 'disabled';
1173
1167
  }
1168
+ /**
1169
+ * Configuration for post-conversation processing. When enabled, the assistant
1170
+ * receives one additional LLM turn after the conversation ends, allowing it to
1171
+ * execute tool calls such as logging to a CRM or sending a summary. The assistant
1172
+ * can execute multiple parallel or sequential tools during this phase.
1173
+ * Telephony-control tools (e.g. hangup, transfer) are unavailable
1174
+ * post-conversation. Beta feature.
1175
+ */
1176
+ export interface PostConversationSettings {
1177
+ /**
1178
+ * Whether post-conversation processing is enabled. When true, the assistant will
1179
+ * be invoked after the conversation ends to perform any final tool calls. Defaults
1180
+ * to false.
1181
+ */
1182
+ enabled?: boolean;
1183
+ }
1184
+ /**
1185
+ * Configuration for post-conversation processing. When enabled, the assistant
1186
+ * receives one additional LLM turn after the conversation ends, allowing it to
1187
+ * execute tool calls such as logging to a CRM or sending a summary. The assistant
1188
+ * can execute multiple parallel or sequential tools during this phase.
1189
+ * Telephony-control tools (e.g. hangup, transfer) are unavailable
1190
+ * post-conversation. Beta feature.
1191
+ */
1192
+ export interface PostConversationSettingsReq {
1193
+ /**
1194
+ * Whether post-conversation processing is enabled. When true, the assistant will
1195
+ * be invoked after the conversation ends to perform any final tool calls. Defaults
1196
+ * to false.
1197
+ */
1198
+ enabled?: boolean;
1199
+ }
1174
1200
  export interface PrivacySettings {
1175
1201
  /**
1176
1202
  * If true, conversation history and insights will be stored. If false, they will
@@ -1772,8 +1798,8 @@ export interface AssistantCreateParams {
1772
1798
  */
1773
1799
  dynamic_variables_webhook_url?: string;
1774
1800
  enabled_features?: Array<EnabledFeatures>;
1775
- external_llm?: AssistantCreateParams.ExternalLlm;
1776
- fallback_config?: AssistantCreateParams.FallbackConfig;
1801
+ external_llm?: ExternalLlmReq;
1802
+ fallback_config?: FallbackConfigReq;
1777
1803
  /**
1778
1804
  * Text that the assistant will use to start the conversation. This may be
1779
1805
  * templated with
@@ -1832,7 +1858,7 @@ export interface AssistantCreateParams {
1832
1858
  * Telephony-control tools (e.g. hangup, transfer) are unavailable
1833
1859
  * post-conversation. Beta feature.
1834
1860
  */
1835
- post_conversation_settings?: AssistantCreateParams.PostConversationSettings;
1861
+ post_conversation_settings?: PostConversationSettingsReq;
1836
1862
  privacy_settings?: PrivacySettings;
1837
1863
  /**
1838
1864
  * Tags associated with the assistant. Tags can also be managed with the assistant
@@ -1859,94 +1885,6 @@ export interface AssistantCreateParams {
1859
1885
  widget_settings?: WidgetSettings;
1860
1886
  }
1861
1887
  export declare namespace AssistantCreateParams {
1862
- interface ExternalLlm {
1863
- /**
1864
- * Base URL for the external LLM endpoint.
1865
- */
1866
- base_url: string;
1867
- /**
1868
- * Model identifier to use with the external LLM endpoint.
1869
- */
1870
- model: string;
1871
- /**
1872
- * Authentication method used when connecting to the external LLM endpoint.
1873
- */
1874
- authentication_method?: 'token' | 'certificate';
1875
- /**
1876
- * Integration secret identifier for the client certificate used with certificate
1877
- * authentication.
1878
- */
1879
- certificate_ref?: string;
1880
- /**
1881
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
1882
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
1883
- * request body. Defaults to `false`. Example payload sent to the external
1884
- * endpoint:
1885
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
1886
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
1887
- * limits.
1888
- */
1889
- forward_metadata?: boolean;
1890
- /**
1891
- * Integration secret identifier for the external LLM API key.
1892
- */
1893
- llm_api_key_ref?: string;
1894
- /**
1895
- * URL used to retrieve an access token when certificate authentication is enabled.
1896
- */
1897
- token_retrieval_url?: string;
1898
- }
1899
- interface FallbackConfig {
1900
- external_llm?: FallbackConfig.ExternalLlm;
1901
- /**
1902
- * Integration secret identifier for the fallback model API key.
1903
- */
1904
- llm_api_key_ref?: string;
1905
- /**
1906
- * Fallback Telnyx-hosted model to use when the primary LLM provider is
1907
- * unavailable.
1908
- */
1909
- model?: string;
1910
- }
1911
- namespace FallbackConfig {
1912
- interface ExternalLlm {
1913
- /**
1914
- * Base URL for the external LLM endpoint.
1915
- */
1916
- base_url: string;
1917
- /**
1918
- * Model identifier to use with the external LLM endpoint.
1919
- */
1920
- model: string;
1921
- /**
1922
- * Authentication method used when connecting to the external LLM endpoint.
1923
- */
1924
- authentication_method?: 'token' | 'certificate';
1925
- /**
1926
- * Integration secret identifier for the client certificate used with certificate
1927
- * authentication.
1928
- */
1929
- certificate_ref?: string;
1930
- /**
1931
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
1932
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
1933
- * request body. Defaults to `false`. Example payload sent to the external
1934
- * endpoint:
1935
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
1936
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
1937
- * limits.
1938
- */
1939
- forward_metadata?: boolean;
1940
- /**
1941
- * Integration secret identifier for the external LLM API key.
1942
- */
1943
- llm_api_key_ref?: string;
1944
- /**
1945
- * URL used to retrieve an access token when certificate authentication is enabled.
1946
- */
1947
- token_retrieval_url?: string;
1948
- }
1949
- }
1950
1888
  /**
1951
1889
  * Reference to a connected integration attached to an assistant. Discover
1952
1890
  * available integrations with `/ai/integrations` and connected integrations with
@@ -2048,22 +1986,6 @@ export declare namespace AssistantCreateParams {
2048
1986
  */
2049
1987
  allowed_tools?: Array<string>;
2050
1988
  }
2051
- /**
2052
- * Configuration for post-conversation processing. When enabled, the assistant
2053
- * receives one additional LLM turn after the conversation ends, allowing it to
2054
- * execute tool calls such as logging to a CRM or sending a summary. The assistant
2055
- * can execute multiple parallel or sequential tools during this phase.
2056
- * Telephony-control tools (e.g. hangup, transfer) are unavailable
2057
- * post-conversation. Beta feature.
2058
- */
2059
- interface PostConversationSettings {
2060
- /**
2061
- * Whether post-conversation processing is enabled. When true, the assistant will
2062
- * be invoked after the conversation ends to perform any final tool calls. Defaults
2063
- * to false.
2064
- */
2065
- enabled?: boolean;
2066
- }
2067
1989
  }
2068
1990
  export interface AssistantRetrieveParams {
2069
1991
  call_control_id?: string;
@@ -2098,8 +2020,8 @@ export interface AssistantUpdateParams {
2098
2020
  */
2099
2021
  dynamic_variables_webhook_url?: string;
2100
2022
  enabled_features?: Array<EnabledFeatures>;
2101
- external_llm?: AssistantUpdateParams.ExternalLlm;
2102
- fallback_config?: AssistantUpdateParams.FallbackConfig;
2023
+ external_llm?: ExternalLlmReq;
2024
+ fallback_config?: FallbackConfigReq;
2103
2025
  /**
2104
2026
  * Text that the assistant will use to start the conversation. This may be
2105
2027
  * templated with
@@ -2164,7 +2086,7 @@ export interface AssistantUpdateParams {
2164
2086
  * Telephony-control tools (e.g. hangup, transfer) are unavailable
2165
2087
  * post-conversation. Beta feature.
2166
2088
  */
2167
- post_conversation_settings?: AssistantUpdateParams.PostConversationSettings;
2089
+ post_conversation_settings?: PostConversationSettingsReq;
2168
2090
  privacy_settings?: PrivacySettings;
2169
2091
  /**
2170
2092
  * Indicates whether the assistant should be promoted to the main version. Defaults
@@ -2200,94 +2122,6 @@ export interface AssistantUpdateParams {
2200
2122
  widget_settings?: WidgetSettings;
2201
2123
  }
2202
2124
  export declare namespace AssistantUpdateParams {
2203
- interface ExternalLlm {
2204
- /**
2205
- * Base URL for the external LLM endpoint.
2206
- */
2207
- base_url: string;
2208
- /**
2209
- * Model identifier to use with the external LLM endpoint.
2210
- */
2211
- model: string;
2212
- /**
2213
- * Authentication method used when connecting to the external LLM endpoint.
2214
- */
2215
- authentication_method?: 'token' | 'certificate';
2216
- /**
2217
- * Integration secret identifier for the client certificate used with certificate
2218
- * authentication.
2219
- */
2220
- certificate_ref?: string;
2221
- /**
2222
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
2223
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
2224
- * request body. Defaults to `false`. Example payload sent to the external
2225
- * endpoint:
2226
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
2227
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
2228
- * limits.
2229
- */
2230
- forward_metadata?: boolean;
2231
- /**
2232
- * Integration secret identifier for the external LLM API key.
2233
- */
2234
- llm_api_key_ref?: string;
2235
- /**
2236
- * URL used to retrieve an access token when certificate authentication is enabled.
2237
- */
2238
- token_retrieval_url?: string;
2239
- }
2240
- interface FallbackConfig {
2241
- external_llm?: FallbackConfig.ExternalLlm;
2242
- /**
2243
- * Integration secret identifier for the fallback model API key.
2244
- */
2245
- llm_api_key_ref?: string;
2246
- /**
2247
- * Fallback Telnyx-hosted model to use when the primary LLM provider is
2248
- * unavailable.
2249
- */
2250
- model?: string;
2251
- }
2252
- namespace FallbackConfig {
2253
- interface ExternalLlm {
2254
- /**
2255
- * Base URL for the external LLM endpoint.
2256
- */
2257
- base_url: string;
2258
- /**
2259
- * Model identifier to use with the external LLM endpoint.
2260
- */
2261
- model: string;
2262
- /**
2263
- * Authentication method used when connecting to the external LLM endpoint.
2264
- */
2265
- authentication_method?: 'token' | 'certificate';
2266
- /**
2267
- * Integration secret identifier for the client certificate used with certificate
2268
- * authentication.
2269
- */
2270
- certificate_ref?: string;
2271
- /**
2272
- * When `true`, Telnyx forwards the assistant's dynamic variables to the external
2273
- * LLM endpoint as a top-level `extra_metadata` object on the chat completion
2274
- * request body. Defaults to `false`. Example payload sent to the external
2275
- * endpoint:
2276
- * `{"extra_metadata": {"customer_name": "Jane", "account_id": "acct_789", "telnyx_agent_target": "+13125550100", "telnyx_end_user_target": "+13125550123"}}`.
2277
- * Distinct from OpenAI's native `metadata` field, which has its own size and type
2278
- * limits.
2279
- */
2280
- forward_metadata?: boolean;
2281
- /**
2282
- * Integration secret identifier for the external LLM API key.
2283
- */
2284
- llm_api_key_ref?: string;
2285
- /**
2286
- * URL used to retrieve an access token when certificate authentication is enabled.
2287
- */
2288
- token_retrieval_url?: string;
2289
- }
2290
- }
2291
2125
  /**
2292
2126
  * Reference to a connected integration attached to an assistant. Discover
2293
2127
  * available integrations with `/ai/integrations` and connected integrations with
@@ -2389,22 +2223,6 @@ export declare namespace AssistantUpdateParams {
2389
2223
  */
2390
2224
  allowed_tools?: Array<string>;
2391
2225
  }
2392
- /**
2393
- * Configuration for post-conversation processing. When enabled, the assistant
2394
- * receives one additional LLM turn after the conversation ends, allowing it to
2395
- * execute tool calls such as logging to a CRM or sending a summary. The assistant
2396
- * can execute multiple parallel or sequential tools during this phase.
2397
- * Telephony-control tools (e.g. hangup, transfer) are unavailable
2398
- * post-conversation. Beta feature.
2399
- */
2400
- interface PostConversationSettings {
2401
- /**
2402
- * Whether post-conversation processing is enabled. When true, the assistant will
2403
- * be invoked after the conversation ends to perform any final tool calls. Defaults
2404
- * to false.
2405
- */
2406
- enabled?: boolean;
2407
- }
2408
2226
  }
2409
2227
  export interface AssistantChatParams {
2410
2228
  /**
@@ -2447,7 +2265,7 @@ export interface AssistantSendSMSParams {
2447
2265
  text?: string;
2448
2266
  }
2449
2267
  export declare namespace Assistants {
2450
- export { type Assistant as Assistant, type AssistantTool as AssistantTool, type AssistantsList as AssistantsList, type AudioVisualizerConfig as AudioVisualizerConfig, type EnabledFeatures as EnabledFeatures, type HangupTool as HangupTool, type HangupToolParams as HangupToolParams, type ImportMetadata as ImportMetadata, type InferenceEmbedding as InferenceEmbedding, type InferenceEmbeddingWebhookToolParams as InferenceEmbeddingWebhookToolParams, type InsightSettings as InsightSettings, type MessagingSettings as MessagingSettings, type Observability as Observability, type ObservabilityReq as ObservabilityReq, type PrivacySettings as PrivacySettings, type RetrievalTool as RetrievalTool, type TelephonySettings as TelephonySettings, type TranscriptionSettings as TranscriptionSettings, type TranscriptionSettingsConfig as TranscriptionSettingsConfig, type TransferTool as TransferTool, type VoiceSettings as VoiceSettings, type WebhookTool as WebhookTool, type WidgetSettings as WidgetSettings, type AssistantDeleteResponse as AssistantDeleteResponse, type AssistantChatResponse as AssistantChatResponse, type AssistantGetTexmlResponse as AssistantGetTexmlResponse, type AssistantSendSMSResponse as AssistantSendSMSResponse, type AssistantCreateParams as AssistantCreateParams, type AssistantRetrieveParams as AssistantRetrieveParams, type AssistantUpdateParams as AssistantUpdateParams, type AssistantChatParams as AssistantChatParams, type AssistantImportsParams as AssistantImportsParams, type AssistantSendSMSParams as AssistantSendSMSParams, };
2268
+ export { type Assistant as Assistant, type AssistantTool as AssistantTool, type AssistantsList as AssistantsList, type AudioVisualizerConfig as AudioVisualizerConfig, type EnabledFeatures as EnabledFeatures, type ExternalLlm as ExternalLlm, type ExternalLlmReq as ExternalLlmReq, type FallbackConfig as FallbackConfig, type FallbackConfigReq as FallbackConfigReq, type HangupTool as HangupTool, type HangupToolParams as HangupToolParams, type ImportMetadata as ImportMetadata, type InferenceEmbedding as InferenceEmbedding, type InferenceEmbeddingWebhookToolParams as InferenceEmbeddingWebhookToolParams, type InsightSettings as InsightSettings, type MessagingSettings as MessagingSettings, type Observability as Observability, type ObservabilityReq as ObservabilityReq, type PostConversationSettings as PostConversationSettings, type PostConversationSettingsReq as PostConversationSettingsReq, type PrivacySettings as PrivacySettings, type RetrievalTool as RetrievalTool, type TelephonySettings as TelephonySettings, type TranscriptionSettings as TranscriptionSettings, type TranscriptionSettingsConfig as TranscriptionSettingsConfig, type TransferTool as TransferTool, type VoiceSettings as VoiceSettings, type WebhookTool as WebhookTool, type WidgetSettings as WidgetSettings, type AssistantDeleteResponse as AssistantDeleteResponse, type AssistantChatResponse as AssistantChatResponse, type AssistantGetTexmlResponse as AssistantGetTexmlResponse, type AssistantSendSMSResponse as AssistantSendSMSResponse, type AssistantCreateParams as AssistantCreateParams, type AssistantRetrieveParams as AssistantRetrieveParams, type AssistantUpdateParams as AssistantUpdateParams, type AssistantChatParams as AssistantChatParams, type AssistantImportsParams as AssistantImportsParams, type AssistantSendSMSParams as AssistantSendSMSParams, };
2451
2269
  export { Tests as Tests, type AssistantTest as AssistantTest, type TelnyxConversationChannel as TelnyxConversationChannel, type AssistantTestsDefaultFlatPagination as AssistantTestsDefaultFlatPagination, type TestCreateParams as TestCreateParams, type TestUpdateParams as TestUpdateParams, type TestListParams as TestListParams, };
2452
2270
  export { CanaryDeploys as CanaryDeploys, type CanaryDeploy as CanaryDeploy, type CanaryDeployResponse as CanaryDeployResponse, type VersionConfig as VersionConfig, type CanaryDeployCreateParams as CanaryDeployCreateParams, type CanaryDeployUpdateParams as CanaryDeployUpdateParams, };
2453
2271
  export { ScheduledEvents as ScheduledEvents, type ConversationChannelType as ConversationChannelType, type EventStatus as EventStatus, type ScheduledEventResponse as ScheduledEventResponse, type ScheduledPhoneCallEventResponse as ScheduledPhoneCallEventResponse, type ScheduledSMSEventResponse as ScheduledSMSEventResponse, type ScheduledEventListResponse as ScheduledEventListResponse, type ScheduledEventListResponsesDefaultFlatPagination as ScheduledEventListResponsesDefaultFlatPagination, type ScheduledEventCreateParams as ScheduledEventCreateParams, type ScheduledEventRetrieveParams as ScheduledEventRetrieveParams, type ScheduledEventListParams as ScheduledEventListParams, type ScheduledEventDeleteParams as ScheduledEventDeleteParams, };