assemblyai 1.0.0.pre.beta.14 → 1.0.0.pre.beta.16

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: b7f6b9de145dc0cb85875b9680ddb0102e0df821868a143ca40dd770003d5932
4
- data.tar.gz: ffc6e2dce5b33b542f25ad316537cbe6bb18032aaaa5e780f955bf0b22f958bd
3
+ metadata.gz: 1d6a5866e5e1540943b02af10a905121eaef30e99c9b53c7c11e7b21d6e65f33
4
+ data.tar.gz: 8eed3a736e9ea9f08e96a48d93394195a2f4c5025f010f965bd50b79b9efb2b7
5
5
  SHA512:
6
- metadata.gz: 15af95d44a789a2d9973a3029aa147b7a869deb2c08988077c2a921ecf7aeab8531d602b4ca101515de253b0a2611f0ef90f99c67bc37432df8bcdaabb03b408
7
- data.tar.gz: 20c104e38452399c68b5b29f3d2e1ab262aadad3c4d1737d80f7a0107fab74ca30a64440620eb570c6da182dc0cd7c64ed32a094ed45e9049ec0d30ae7140b04
6
+ metadata.gz: 6bca4577ee221e5b131f331fd9527c0bb61fcb481ce8dcf154562f2e54a20566eccb0a4d150680d44397ecf2c760b3b38d22b3a4e9093fb4e814d740de0f5baf
7
+ data.tar.gz: 306697c4cb1583b4410e60c6a49b86d01c2a2486dd121ec43d1bb16da93cc4813ca407f4bad19488680bc2034d1216879da0c5174d9ebe9991a716038ef99d1b
@@ -33,7 +33,6 @@ module AssemblyAI
33
33
  # Use either transcript_ids or input_text as input into LeMUR.
34
34
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
35
35
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
36
- # Defaults to "default".
37
36
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
38
37
  # @param temperature [Float] The temperature to use for the model.
39
38
  # Higher values result in answers that are more creative, lower values are more
@@ -84,7 +83,6 @@ module AssemblyAI
84
83
  # Use either transcript_ids or input_text as input into LeMUR.
85
84
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
86
85
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
87
- # Defaults to "default".
88
86
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
89
87
  # @param temperature [Float] The temperature to use for the model.
90
88
  # Higher values result in answers that are more creative, lower values are more
@@ -136,7 +134,6 @@ module AssemblyAI
136
134
  # Use either transcript_ids or input_text as input into LeMUR.
137
135
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
138
136
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
139
- # Defaults to "default".
140
137
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
141
138
  # @param temperature [Float] The temperature to use for the model.
142
139
  # Higher values result in answers that are more creative, lower values are more
@@ -187,7 +184,6 @@ module AssemblyAI
187
184
  # Use either transcript_ids or input_text as input into LeMUR.
188
185
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
189
186
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
190
- # Defaults to "default".
191
187
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
192
188
  # @param temperature [Float] The temperature to use for the model.
193
189
  # Higher values result in answers that are more creative, lower values are more
@@ -294,7 +290,6 @@ module AssemblyAI
294
290
  # Use either transcript_ids or input_text as input into LeMUR.
295
291
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
296
292
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
297
- # Defaults to "default".
298
293
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
299
294
  # @param temperature [Float] The temperature to use for the model.
300
295
  # Higher values result in answers that are more creative, lower values are more
@@ -347,7 +342,6 @@ module AssemblyAI
347
342
  # Use either transcript_ids or input_text as input into LeMUR.
348
343
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
349
344
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
350
- # Defaults to "default".
351
345
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
352
346
  # @param temperature [Float] The temperature to use for the model.
353
347
  # Higher values result in answers that are more creative, lower values are more
@@ -401,7 +395,6 @@ module AssemblyAI
401
395
  # Use either transcript_ids or input_text as input into LeMUR.
402
396
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
403
397
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
404
- # Defaults to "default".
405
398
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
406
399
  # @param temperature [Float] The temperature to use for the model.
407
400
  # Higher values result in answers that are more creative, lower values are more
@@ -454,7 +447,6 @@ module AssemblyAI
454
447
  # Use either transcript_ids or input_text as input into LeMUR.
455
448
  # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
456
449
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
457
- # Defaults to "default".
458
450
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
459
451
  # @param temperature [Float] The temperature to use for the model.
460
452
  # Higher values result in answers that are more creative, lower values are more
@@ -19,7 +19,6 @@ module AssemblyAI
19
19
  # @return [AssemblyAI::Lemur::LemurBaseParamsContext] Context to provide the model. This can be a string or a free-form JSON value.
20
20
  attr_reader :context
21
21
  # @return [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
22
- # Defaults to "default".
23
22
  attr_reader :final_model
24
23
  # @return [Integer] Max output size in tokens, up to 4000
25
24
  attr_reader :max_output_size
@@ -44,7 +43,6 @@ module AssemblyAI
44
43
  # Use either transcript_ids or input_text as input into LeMUR.
45
44
  # @param context [AssemblyAI::Lemur::LemurBaseParamsContext] Context to provide the model. This can be a string or a free-form JSON value.
46
45
  # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
47
- # Defaults to "default".
48
46
  # @param max_output_size [Integer] Max output size in tokens, up to 4000
49
47
  # @param temperature [Float] The temperature to use for the model.
50
48
  # Higher values result in answers that are more creative, lower values are more
@@ -4,10 +4,16 @@ module AssemblyAI
4
4
  class Lemur
5
5
  # The model that is used for the final prompt after compression is performed.
6
6
  class LemurModel
7
+ ANTHROPIC_CLAUDE3_5_SONNET = "anthropic/claude-3-5-sonnet"
8
+ ANTHROPIC_CLAUDE3_OPUS = "anthropic/claude-3-opus"
9
+ ANTHROPIC_CLAUDE3_HAIKU = "anthropic/claude-3-haiku"
10
+ ANTHROPIC_CLAUDE3_SONNET = "anthropic/claude-3-sonnet"
11
+ ANTHROPIC_CLAUDE2_1 = "anthropic/claude-2-1"
12
+ ANTHROPIC_CLAUDE2 = "anthropic/claude-2"
7
13
  DEFAULT = "default"
14
+ ANTHROPIC_CLAUDE_INSTANT1_2 = "anthropic/claude-instant-1-2"
8
15
  BASIC = "basic"
9
16
  ASSEMBLYAI_MISTRAL7B = "assemblyai/mistral-7b"
10
- ANTHROPIC_CLAUDE2_1 = "anthropic/claude-2-1"
11
17
  end
12
18
  end
13
19
  end
data/lib/gemconfig.rb CHANGED
@@ -2,7 +2,7 @@
2
2
 
3
3
  module AssemblyAI
4
4
  module Gemconfig
5
- VERSION = "1.0.0-beta.14"
5
+ VERSION = "1.0.0-beta.16"
6
6
  AUTHORS = [""].freeze
7
7
  EMAIL = "support@assemblyai.com"
8
8
  SUMMARY = "AssemblyAI Ruby SDK"
data/lib/requests.rb CHANGED
@@ -2,6 +2,7 @@
2
2
 
3
3
  require_relative "environment"
4
4
  require_relative "user_agent"
5
+ require_relative "gemconfig"
5
6
  require "faraday"
6
7
  require "faraday/retry"
7
8
  require "async/http/faraday"
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: assemblyai
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.0.0.pre.beta.14
4
+ version: 1.0.0.pre.beta.16
5
5
  platform: ruby
6
6
  authors:
7
7
  - ''
8
8
  autorequire:
9
9
  bindir: exe
10
10
  cert_chain: []
11
- date: 2024-07-04 00:00:00.000000000 Z
11
+ date: 2024-07-15 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: async-http-faraday