edsl 0.1.39.dev1__py3-none-any.whl → 0.1.39.dev3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (256) hide show
  1. edsl/Base.py +332 -332
  2. edsl/BaseDiff.py +260 -260
  3. edsl/TemplateLoader.py +24 -24
  4. edsl/__init__.py +49 -49
  5. edsl/__version__.py +1 -1
  6. edsl/agents/Agent.py +867 -867
  7. edsl/agents/AgentList.py +413 -413
  8. edsl/agents/Invigilator.py +233 -233
  9. edsl/agents/InvigilatorBase.py +270 -265
  10. edsl/agents/PromptConstructor.py +354 -354
  11. edsl/agents/__init__.py +3 -3
  12. edsl/agents/descriptors.py +99 -99
  13. edsl/agents/prompt_helpers.py +129 -129
  14. edsl/auto/AutoStudy.py +117 -117
  15. edsl/auto/StageBase.py +230 -230
  16. edsl/auto/StageGenerateSurvey.py +178 -178
  17. edsl/auto/StageLabelQuestions.py +125 -125
  18. edsl/auto/StagePersona.py +61 -61
  19. edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
  20. edsl/auto/StagePersonaDimensionValues.py +74 -74
  21. edsl/auto/StagePersonaDimensions.py +69 -69
  22. edsl/auto/StageQuestions.py +73 -73
  23. edsl/auto/SurveyCreatorPipeline.py +21 -21
  24. edsl/auto/utilities.py +224 -224
  25. edsl/base/Base.py +279 -279
  26. edsl/config.py +157 -157
  27. edsl/conversation/Conversation.py +290 -290
  28. edsl/conversation/car_buying.py +58 -58
  29. edsl/conversation/chips.py +95 -95
  30. edsl/conversation/mug_negotiation.py +81 -81
  31. edsl/conversation/next_speaker_utilities.py +93 -93
  32. edsl/coop/PriceFetcher.py +54 -54
  33. edsl/coop/__init__.py +2 -2
  34. edsl/coop/coop.py +1028 -1028
  35. edsl/coop/utils.py +131 -131
  36. edsl/data/Cache.py +555 -555
  37. edsl/data/CacheEntry.py +233 -233
  38. edsl/data/CacheHandler.py +149 -149
  39. edsl/data/RemoteCacheSync.py +78 -78
  40. edsl/data/SQLiteDict.py +292 -292
  41. edsl/data/__init__.py +4 -4
  42. edsl/data/orm.py +10 -10
  43. edsl/data_transfer_models.py +73 -73
  44. edsl/enums.py +175 -175
  45. edsl/exceptions/BaseException.py +21 -21
  46. edsl/exceptions/__init__.py +54 -54
  47. edsl/exceptions/agents.py +42 -42
  48. edsl/exceptions/cache.py +5 -5
  49. edsl/exceptions/configuration.py +16 -16
  50. edsl/exceptions/coop.py +10 -10
  51. edsl/exceptions/data.py +14 -14
  52. edsl/exceptions/general.py +34 -34
  53. edsl/exceptions/jobs.py +33 -33
  54. edsl/exceptions/language_models.py +63 -63
  55. edsl/exceptions/prompts.py +15 -15
  56. edsl/exceptions/questions.py +91 -91
  57. edsl/exceptions/results.py +29 -29
  58. edsl/exceptions/scenarios.py +22 -22
  59. edsl/exceptions/surveys.py +37 -37
  60. edsl/inference_services/AnthropicService.py +87 -87
  61. edsl/inference_services/AwsBedrock.py +120 -120
  62. edsl/inference_services/AzureAI.py +217 -217
  63. edsl/inference_services/DeepInfraService.py +18 -18
  64. edsl/inference_services/GoogleService.py +148 -148
  65. edsl/inference_services/GroqService.py +20 -20
  66. edsl/inference_services/InferenceServiceABC.py +147 -147
  67. edsl/inference_services/InferenceServicesCollection.py +97 -97
  68. edsl/inference_services/MistralAIService.py +123 -123
  69. edsl/inference_services/OllamaService.py +18 -18
  70. edsl/inference_services/OpenAIService.py +224 -224
  71. edsl/inference_services/PerplexityService.py +163 -163
  72. edsl/inference_services/TestService.py +89 -89
  73. edsl/inference_services/TogetherAIService.py +170 -170
  74. edsl/inference_services/models_available_cache.py +118 -118
  75. edsl/inference_services/rate_limits_cache.py +25 -25
  76. edsl/inference_services/registry.py +41 -41
  77. edsl/inference_services/write_available.py +10 -10
  78. edsl/jobs/Answers.py +56 -56
  79. edsl/jobs/Jobs.py +898 -898
  80. edsl/jobs/JobsChecks.py +147 -147
  81. edsl/jobs/JobsPrompts.py +268 -268
  82. edsl/jobs/JobsRemoteInferenceHandler.py +239 -239
  83. edsl/jobs/__init__.py +1 -1
  84. edsl/jobs/buckets/BucketCollection.py +63 -63
  85. edsl/jobs/buckets/ModelBuckets.py +65 -65
  86. edsl/jobs/buckets/TokenBucket.py +251 -251
  87. edsl/jobs/interviews/Interview.py +661 -661
  88. edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
  89. edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
  90. edsl/jobs/interviews/InterviewStatistic.py +63 -63
  91. edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
  92. edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
  93. edsl/jobs/interviews/InterviewStatusLog.py +92 -92
  94. edsl/jobs/interviews/ReportErrors.py +66 -66
  95. edsl/jobs/interviews/interview_status_enum.py +9 -9
  96. edsl/jobs/runners/JobsRunnerAsyncio.py +466 -466
  97. edsl/jobs/runners/JobsRunnerStatus.py +330 -330
  98. edsl/jobs/tasks/QuestionTaskCreator.py +242 -242
  99. edsl/jobs/tasks/TaskCreators.py +64 -64
  100. edsl/jobs/tasks/TaskHistory.py +450 -450
  101. edsl/jobs/tasks/TaskStatusLog.py +23 -23
  102. edsl/jobs/tasks/task_status_enum.py +163 -163
  103. edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
  104. edsl/jobs/tokens/TokenUsage.py +34 -34
  105. edsl/language_models/KeyLookup.py +30 -30
  106. edsl/language_models/LanguageModel.py +668 -668
  107. edsl/language_models/ModelList.py +155 -155
  108. edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
  109. edsl/language_models/__init__.py +3 -3
  110. edsl/language_models/fake_openai_call.py +15 -15
  111. edsl/language_models/fake_openai_service.py +61 -61
  112. edsl/language_models/registry.py +190 -190
  113. edsl/language_models/repair.py +156 -156
  114. edsl/language_models/unused/ReplicateBase.py +83 -83
  115. edsl/language_models/utilities.py +64 -64
  116. edsl/notebooks/Notebook.py +258 -258
  117. edsl/notebooks/__init__.py +1 -1
  118. edsl/prompts/Prompt.py +362 -362
  119. edsl/prompts/__init__.py +2 -2
  120. edsl/questions/AnswerValidatorMixin.py +289 -289
  121. edsl/questions/QuestionBase.py +664 -664
  122. edsl/questions/QuestionBaseGenMixin.py +161 -161
  123. edsl/questions/QuestionBasePromptsMixin.py +217 -217
  124. edsl/questions/QuestionBudget.py +227 -227
  125. edsl/questions/QuestionCheckBox.py +359 -359
  126. edsl/questions/QuestionExtract.py +182 -182
  127. edsl/questions/QuestionFreeText.py +114 -114
  128. edsl/questions/QuestionFunctional.py +166 -166
  129. edsl/questions/QuestionList.py +231 -231
  130. edsl/questions/QuestionMultipleChoice.py +286 -286
  131. edsl/questions/QuestionNumerical.py +153 -153
  132. edsl/questions/QuestionRank.py +324 -324
  133. edsl/questions/Quick.py +41 -41
  134. edsl/questions/RegisterQuestionsMeta.py +71 -71
  135. edsl/questions/ResponseValidatorABC.py +174 -174
  136. edsl/questions/SimpleAskMixin.py +73 -73
  137. edsl/questions/__init__.py +26 -26
  138. edsl/questions/compose_questions.py +98 -98
  139. edsl/questions/decorators.py +21 -21
  140. edsl/questions/derived/QuestionLikertFive.py +76 -76
  141. edsl/questions/derived/QuestionLinearScale.py +87 -87
  142. edsl/questions/derived/QuestionTopK.py +93 -93
  143. edsl/questions/derived/QuestionYesNo.py +82 -82
  144. edsl/questions/descriptors.py +413 -413
  145. edsl/questions/prompt_templates/question_budget.jinja +13 -13
  146. edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
  147. edsl/questions/prompt_templates/question_extract.jinja +11 -11
  148. edsl/questions/prompt_templates/question_free_text.jinja +3 -3
  149. edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
  150. edsl/questions/prompt_templates/question_list.jinja +17 -17
  151. edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
  152. edsl/questions/prompt_templates/question_numerical.jinja +36 -36
  153. edsl/questions/question_registry.py +177 -177
  154. edsl/questions/settings.py +12 -12
  155. edsl/questions/templates/budget/answering_instructions.jinja +7 -7
  156. edsl/questions/templates/budget/question_presentation.jinja +7 -7
  157. edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
  158. edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
  159. edsl/questions/templates/extract/answering_instructions.jinja +7 -7
  160. edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
  161. edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
  162. edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
  163. edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
  164. edsl/questions/templates/list/answering_instructions.jinja +3 -3
  165. edsl/questions/templates/list/question_presentation.jinja +5 -5
  166. edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
  167. edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
  168. edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
  169. edsl/questions/templates/numerical/question_presentation.jinja +6 -6
  170. edsl/questions/templates/rank/answering_instructions.jinja +11 -11
  171. edsl/questions/templates/rank/question_presentation.jinja +15 -15
  172. edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
  173. edsl/questions/templates/top_k/question_presentation.jinja +22 -22
  174. edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
  175. edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
  176. edsl/results/CSSParameterizer.py +108 -108
  177. edsl/results/Dataset.py +424 -424
  178. edsl/results/DatasetExportMixin.py +731 -731
  179. edsl/results/DatasetTree.py +275 -275
  180. edsl/results/Result.py +465 -465
  181. edsl/results/Results.py +1165 -1165
  182. edsl/results/ResultsDBMixin.py +238 -238
  183. edsl/results/ResultsExportMixin.py +43 -43
  184. edsl/results/ResultsFetchMixin.py +33 -33
  185. edsl/results/ResultsGGMixin.py +121 -121
  186. edsl/results/ResultsToolsMixin.py +98 -98
  187. edsl/results/Selector.py +135 -135
  188. edsl/results/TableDisplay.py +198 -198
  189. edsl/results/__init__.py +2 -2
  190. edsl/results/table_display.css +77 -77
  191. edsl/results/tree_explore.py +115 -115
  192. edsl/scenarios/FileStore.py +632 -632
  193. edsl/scenarios/Scenario.py +601 -601
  194. edsl/scenarios/ScenarioHtmlMixin.py +64 -64
  195. edsl/scenarios/ScenarioJoin.py +127 -127
  196. edsl/scenarios/ScenarioList.py +1287 -1287
  197. edsl/scenarios/ScenarioListExportMixin.py +52 -52
  198. edsl/scenarios/ScenarioListPdfMixin.py +261 -261
  199. edsl/scenarios/__init__.py +4 -4
  200. edsl/shared.py +1 -1
  201. edsl/study/ObjectEntry.py +173 -173
  202. edsl/study/ProofOfWork.py +113 -113
  203. edsl/study/SnapShot.py +80 -80
  204. edsl/study/Study.py +528 -528
  205. edsl/study/__init__.py +4 -4
  206. edsl/surveys/DAG.py +148 -148
  207. edsl/surveys/Memory.py +31 -31
  208. edsl/surveys/MemoryPlan.py +244 -244
  209. edsl/surveys/Rule.py +326 -326
  210. edsl/surveys/RuleCollection.py +387 -387
  211. edsl/surveys/Survey.py +1801 -1801
  212. edsl/surveys/SurveyCSS.py +261 -261
  213. edsl/surveys/SurveyExportMixin.py +259 -259
  214. edsl/surveys/SurveyFlowVisualizationMixin.py +179 -179
  215. edsl/surveys/SurveyQualtricsImport.py +284 -284
  216. edsl/surveys/__init__.py +3 -3
  217. edsl/surveys/base.py +53 -53
  218. edsl/surveys/descriptors.py +56 -56
  219. edsl/surveys/instructions/ChangeInstruction.py +49 -49
  220. edsl/surveys/instructions/Instruction.py +65 -65
  221. edsl/surveys/instructions/InstructionCollection.py +77 -77
  222. edsl/templates/error_reporting/base.html +23 -23
  223. edsl/templates/error_reporting/exceptions_by_model.html +34 -34
  224. edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
  225. edsl/templates/error_reporting/exceptions_by_type.html +16 -16
  226. edsl/templates/error_reporting/interview_details.html +115 -115
  227. edsl/templates/error_reporting/interviews.html +19 -19
  228. edsl/templates/error_reporting/overview.html +4 -4
  229. edsl/templates/error_reporting/performance_plot.html +1 -1
  230. edsl/templates/error_reporting/report.css +73 -73
  231. edsl/templates/error_reporting/report.html +117 -117
  232. edsl/templates/error_reporting/report.js +25 -25
  233. edsl/tools/__init__.py +1 -1
  234. edsl/tools/clusters.py +192 -192
  235. edsl/tools/embeddings.py +27 -27
  236. edsl/tools/embeddings_plotting.py +118 -118
  237. edsl/tools/plotting.py +112 -112
  238. edsl/tools/summarize.py +18 -18
  239. edsl/utilities/SystemInfo.py +28 -28
  240. edsl/utilities/__init__.py +22 -22
  241. edsl/utilities/ast_utilities.py +25 -25
  242. edsl/utilities/data/Registry.py +6 -6
  243. edsl/utilities/data/__init__.py +1 -1
  244. edsl/utilities/data/scooter_results.json +1 -1
  245. edsl/utilities/decorators.py +77 -77
  246. edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
  247. edsl/utilities/interface.py +627 -627
  248. edsl/utilities/naming_utilities.py +263 -263
  249. edsl/utilities/repair_functions.py +28 -28
  250. edsl/utilities/restricted_python.py +70 -70
  251. edsl/utilities/utilities.py +424 -424
  252. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/LICENSE +21 -21
  253. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/METADATA +1 -1
  254. edsl-0.1.39.dev3.dist-info/RECORD +277 -0
  255. edsl-0.1.39.dev1.dist-info/RECORD +0 -277
  256. {edsl-0.1.39.dev1.dist-info → edsl-0.1.39.dev3.dist-info}/WHEEL +0 -0
@@ -1,65 +1,65 @@
1
- # from edsl.jobs.buckets.TokenBucket import TokenBucket
2
-
3
-
4
- class ModelBuckets:
5
- """A class to represent the token and request buckets for a model.
6
-
7
- Most LLM model services have limits both on requests-per-minute (RPM) and tokens-per-minute (TPM).
8
- A request is one call to the service. The number of tokens required for a request depends on parameters.
9
- """
10
-
11
- def __init__(self, requests_bucket: "TokenBucket", tokens_bucket: "TokenBucket"):
12
- """Initialize the model buckets.
13
-
14
- The requests bucket captures requests per unit of time.
15
- The tokens bucket captures the number of language model tokens.
16
-
17
- """
18
- self.requests_bucket = requests_bucket
19
- self.tokens_bucket = tokens_bucket
20
-
21
- def __add__(self, other: "ModelBuckets"):
22
- """Combine two model buckets."""
23
- return ModelBuckets(
24
- requests_bucket=self.requests_bucket + other.requests_bucket,
25
- tokens_bucket=self.tokens_bucket + other.tokens_bucket,
26
- )
27
-
28
- def turbo_mode_on(self):
29
- """Set the refill rate to infinity for both buckets."""
30
- self.requests_bucket.turbo_mode_on()
31
- self.tokens_bucket.turbo_mode_on()
32
-
33
- def turbo_mode_off(self):
34
- """Restore the refill rate to its original value for both buckets."""
35
- self.requests_bucket.turbo_mode_off()
36
- self.tokens_bucket.turbo_mode_off()
37
-
38
- @classmethod
39
- def infinity_bucket(cls, model_name: str = "not_specified") -> "ModelBuckets":
40
- """Create a bucket with infinite capacity and refill rate."""
41
- from edsl.jobs.buckets.TokenBucket import TokenBucket
42
-
43
- return cls(
44
- requests_bucket=TokenBucket(
45
- bucket_name=model_name,
46
- bucket_type="requests",
47
- capacity=float("inf"),
48
- refill_rate=float("inf"),
49
- ),
50
- tokens_bucket=TokenBucket(
51
- bucket_name=model_name,
52
- bucket_type="tokens",
53
- capacity=float("inf"),
54
- refill_rate=float("inf"),
55
- ),
56
- )
57
-
58
- def visualize(self):
59
- """Visualize the token and request buckets."""
60
- plot1 = self.requests_bucket.visualize()
61
- plot2 = self.tokens_bucket.visualize()
62
- return plot1, plot2
63
-
64
- def __repr__(self):
65
- return f"ModelBuckets(requests_bucket={self.requests_bucket}, tokens_bucket={self.tokens_bucket})"
1
+ # from edsl.jobs.buckets.TokenBucket import TokenBucket
2
+
3
+
4
+ class ModelBuckets:
5
+ """A class to represent the token and request buckets for a model.
6
+
7
+ Most LLM model services have limits both on requests-per-minute (RPM) and tokens-per-minute (TPM).
8
+ A request is one call to the service. The number of tokens required for a request depends on parameters.
9
+ """
10
+
11
+ def __init__(self, requests_bucket: "TokenBucket", tokens_bucket: "TokenBucket"):
12
+ """Initialize the model buckets.
13
+
14
+ The requests bucket captures requests per unit of time.
15
+ The tokens bucket captures the number of language model tokens.
16
+
17
+ """
18
+ self.requests_bucket = requests_bucket
19
+ self.tokens_bucket = tokens_bucket
20
+
21
+ def __add__(self, other: "ModelBuckets"):
22
+ """Combine two model buckets."""
23
+ return ModelBuckets(
24
+ requests_bucket=self.requests_bucket + other.requests_bucket,
25
+ tokens_bucket=self.tokens_bucket + other.tokens_bucket,
26
+ )
27
+
28
+ def turbo_mode_on(self):
29
+ """Set the refill rate to infinity for both buckets."""
30
+ self.requests_bucket.turbo_mode_on()
31
+ self.tokens_bucket.turbo_mode_on()
32
+
33
+ def turbo_mode_off(self):
34
+ """Restore the refill rate to its original value for both buckets."""
35
+ self.requests_bucket.turbo_mode_off()
36
+ self.tokens_bucket.turbo_mode_off()
37
+
38
+ @classmethod
39
+ def infinity_bucket(cls, model_name: str = "not_specified") -> "ModelBuckets":
40
+ """Create a bucket with infinite capacity and refill rate."""
41
+ from edsl.jobs.buckets.TokenBucket import TokenBucket
42
+
43
+ return cls(
44
+ requests_bucket=TokenBucket(
45
+ bucket_name=model_name,
46
+ bucket_type="requests",
47
+ capacity=float("inf"),
48
+ refill_rate=float("inf"),
49
+ ),
50
+ tokens_bucket=TokenBucket(
51
+ bucket_name=model_name,
52
+ bucket_type="tokens",
53
+ capacity=float("inf"),
54
+ refill_rate=float("inf"),
55
+ ),
56
+ )
57
+
58
+ def visualize(self):
59
+ """Visualize the token and request buckets."""
60
+ plot1 = self.requests_bucket.visualize()
61
+ plot2 = self.tokens_bucket.visualize()
62
+ return plot1, plot2
63
+
64
+ def __repr__(self):
65
+ return f"ModelBuckets(requests_bucket={self.requests_bucket}, tokens_bucket={self.tokens_bucket})"
@@ -1,251 +1,251 @@
1
- from typing import Union, List, Any, Optional
2
- import asyncio
3
- import time
4
-
5
-
6
- class TokenBucket:
7
- """This is a token bucket used to respect rate limits to services."""
8
-
9
- def __init__(
10
- self,
11
- *,
12
- bucket_name: str,
13
- bucket_type: str,
14
- capacity: Union[int, float],
15
- refill_rate: Union[int, float],
16
- ):
17
- self.bucket_name = bucket_name
18
- self.bucket_type = bucket_type
19
- self.capacity = capacity # Maximum number of tokens
20
- self.added_tokens = 0
21
-
22
- self.target_rate = (
23
- capacity * 60
24
- ) # set this here because it can change with turbo mode
25
-
26
- self._old_capacity = capacity
27
- self.tokens = capacity # Current number of available tokens
28
- self.refill_rate = refill_rate # Rate at which tokens are refilled
29
- self._old_refill_rate = refill_rate
30
- self.last_refill = time.monotonic() # Last refill time
31
- self.log: List[Any] = []
32
- self.turbo_mode = False
33
-
34
- self.creation_time = time.monotonic()
35
-
36
- self.num_requests = 0
37
- self.num_released = 0
38
- self.tokens_returned = 0
39
-
40
- def turbo_mode_on(self):
41
- """Set the refill rate to infinity."""
42
- if self.turbo_mode:
43
- pass
44
- else:
45
- # pass
46
- self.turbo_mode = True
47
- self.capacity = float("inf")
48
- self.refill_rate = float("inf")
49
-
50
- def turbo_mode_off(self):
51
- """Restore the refill rate to its original value."""
52
- self.turbo_mode = False
53
- self.capacity = self._old_capacity
54
- self.refill_rate = self._old_refill_rate
55
-
56
- def __add__(self, other) -> "TokenBucket":
57
- """Combine two token buckets.
58
-
59
- The resulting bucket has the minimum capacity and refill rate of the two buckets.
60
- This is useful, for example, if we have two calls to the same model on the same service but have different temperatures.
61
- """
62
- return TokenBucket(
63
- bucket_name=self.bucket_name,
64
- bucket_type=self.bucket_type,
65
- capacity=min(self.capacity, other.capacity),
66
- refill_rate=min(self.refill_rate, other.refill_rate),
67
- )
68
-
69
- def __repr__(self):
70
- return f"TokenBucket(bucket_name={self.bucket_name}, bucket_type='{self.bucket_type}', capacity={self.capacity}, refill_rate={self.refill_rate})"
71
-
72
- def add_tokens(self, tokens: Union[int, float]) -> None:
73
- """Add tokens to the bucket, up to the maximum capacity.
74
-
75
- :param tokens: The number of tokens to add to the bucket.
76
-
77
- >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
78
- >>> bucket.tokens
79
- 10
80
- >>> bucket.add_tokens(5)
81
- >>> bucket.tokens
82
- 10
83
- """
84
- self.tokens_returned += tokens
85
- self.tokens = min(self.capacity, self.tokens + tokens)
86
- self.log.append((time.monotonic(), self.tokens))
87
-
88
- def refill(self) -> None:
89
- """Refill the bucket with new tokens based on elapsed time.
90
-
91
-
92
-
93
- >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
94
- >>> bucket.tokens = 0
95
- >>> bucket.refill()
96
- >>> bucket.tokens > 0
97
- True
98
- """
99
- """Refill the bucket with new tokens based on elapsed time."""
100
- now = time.monotonic()
101
- # print(f"Time is now: {now}; Last refill time: {self.last_refill}")
102
- elapsed = now - self.last_refill
103
- # print("Elapsed time: ", elapsed)
104
- refill_amount = elapsed * self.refill_rate
105
- self.tokens = min(self.capacity, self.tokens + refill_amount)
106
- self.last_refill = now
107
-
108
- if self.tokens < self.capacity:
109
- pass
110
- # print(f"Refilled. Current tokens: {self.tokens:.4f}")
111
- # print(f"Elapsed time: {elapsed:.4f} seconds")
112
- # print(f"Refill amount: {refill_amount:.4f}")
113
-
114
- self.log.append((now, self.tokens))
115
-
116
- def wait_time(self, requested_tokens: Union[float, int]) -> float:
117
- """Calculate the time to wait for the requested number of tokens."""
118
- # self.refill() # Update the current token count
119
- if self.tokens >= requested_tokens:
120
- return 0
121
- return (requested_tokens - self.tokens) / self.refill_rate
122
-
123
- async def get_tokens(
124
- self, amount: Union[int, float] = 1, cheat_bucket_capacity=True
125
- ) -> None:
126
- """Wait for the specified number of tokens to become available.
127
-
128
-
129
- :param amount: The number of tokens
130
- :param warn: If True, warn if the requested amount exceeds the bucket capacity.
131
-
132
- >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
133
- >>> asyncio.run(bucket.get_tokens(5))
134
- >>> bucket.tokens
135
- 5
136
- >>> asyncio.run(bucket.get_tokens(9))
137
- >>> bucket.tokens < 1
138
- True
139
-
140
- >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
141
- >>> asyncio.run(bucket.get_tokens(11, cheat_bucket_capacity=False))
142
- Traceback (most recent call last):
143
- ...
144
- ValueError: Requested amount exceeds bucket capacity. Bucket capacity: 10, requested amount: 11. As the bucket never overflows, the requested amount will never be available.
145
- >>> asyncio.run(bucket.get_tokens(11, cheat_bucket_capacity=True))
146
- >>> bucket.capacity
147
- 12.100000000000001
148
- """
149
- self.num_requests += amount
150
- if amount >= self.capacity:
151
- if not cheat_bucket_capacity:
152
- msg = f"Requested amount exceeds bucket capacity. Bucket capacity: {self.capacity}, requested amount: {amount}. As the bucket never overflows, the requested amount will never be available."
153
- raise ValueError(msg)
154
- else:
155
- self.capacity = amount * 1.10
156
- self._old_capacity = self.capacity
157
-
158
- start_time = time.monotonic()
159
- while True:
160
- self.refill() # Refill based on elapsed time
161
- if self.tokens >= amount:
162
- self.tokens -= amount
163
- break
164
-
165
- wait_time = self.wait_time(amount)
166
- if wait_time > 0:
167
- await asyncio.sleep(wait_time)
168
-
169
- self.num_released += amount
170
- now = time.monotonic()
171
- self.log.append((now, self.tokens))
172
- return None
173
-
174
- def get_log(self) -> list[tuple]:
175
- return self.log
176
-
177
- def visualize(self):
178
- """Visualize the token bucket over time."""
179
- times, tokens = zip(*self.get_log())
180
- start_time = times[0]
181
- times = [t - start_time for t in times] # Normalize time to start from 0
182
- from matplotlib import pyplot as plt
183
-
184
- plt.figure(figsize=(10, 6))
185
- plt.plot(times, tokens, label="Tokens Available")
186
- plt.xlabel("Time (seconds)", fontsize=12)
187
- plt.ylabel("Number of Tokens", fontsize=12)
188
- details = f"{self.bucket_name} ({self.bucket_type}) Bucket Usage Over Time\nCapacity: {self.capacity:.1f}, Refill Rate: {self.refill_rate:.1f}/second"
189
- plt.title(details, fontsize=14)
190
-
191
- plt.legend()
192
- plt.grid(True)
193
- plt.tight_layout()
194
- plt.show()
195
-
196
- def get_throughput(self, time_window: Optional[float] = None) -> float:
197
- """
198
- Calculate the empirical bucket throughput in tokens per minute for the specified time window.
199
-
200
- :param time_window: The time window in seconds to calculate the throughput for.
201
- :return: The throughput in tokens per minute.
202
-
203
- >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=100, refill_rate=10)
204
- >>> asyncio.run(bucket.get_tokens(50))
205
- >>> time.sleep(1) # Wait for 1 second
206
- >>> asyncio.run(bucket.get_tokens(30))
207
- >>> throughput = bucket.get_throughput(1)
208
- >>> 4750 < throughput < 4850
209
- True
210
- """
211
- now = time.monotonic()
212
-
213
- if time_window is None:
214
- start_time = self.creation_time
215
- else:
216
- start_time = now - time_window
217
-
218
- if start_time < self.creation_time:
219
- start_time = self.creation_time
220
-
221
- elapsed_time = now - start_time
222
-
223
- if elapsed_time == 0:
224
- return self.num_released / 0.001
225
-
226
- return (self.num_released / elapsed_time) * 60
227
-
228
- # # Filter log entries within the time window
229
- # relevant_log = [(t, tokens) for t, tokens in self.log if t >= start_time]
230
-
231
- # if len(relevant_log) < 2:
232
- # return 0 # Not enough data points to calculate throughput
233
-
234
- # # Calculate total tokens used
235
- # initial_tokens = relevant_log[0][1]
236
- # final_tokens = relevant_log[-1][1]
237
- # tokens_used = self.num_released - (final_tokens - initial_tokens)
238
-
239
- # # Calculate actual time elapsed
240
- # actual_time_elapsed = relevant_log[-1][0] - relevant_log[0][0]
241
-
242
- # # Calculate throughput in tokens per minute
243
- # throughput = (tokens_used / actual_time_elapsed) * 60
244
-
245
- # return throughput
246
-
247
-
248
- if __name__ == "__main__":
249
- import doctest
250
-
251
- doctest.testmod(optionflags=doctest.ELLIPSIS)
1
+ from typing import Union, List, Any, Optional
2
+ import asyncio
3
+ import time
4
+
5
+
6
+ class TokenBucket:
7
+ """This is a token bucket used to respect rate limits to services."""
8
+
9
+ def __init__(
10
+ self,
11
+ *,
12
+ bucket_name: str,
13
+ bucket_type: str,
14
+ capacity: Union[int, float],
15
+ refill_rate: Union[int, float],
16
+ ):
17
+ self.bucket_name = bucket_name
18
+ self.bucket_type = bucket_type
19
+ self.capacity = capacity # Maximum number of tokens
20
+ self.added_tokens = 0
21
+
22
+ self.target_rate = (
23
+ capacity * 60
24
+ ) # set this here because it can change with turbo mode
25
+
26
+ self._old_capacity = capacity
27
+ self.tokens = capacity # Current number of available tokens
28
+ self.refill_rate = refill_rate # Rate at which tokens are refilled
29
+ self._old_refill_rate = refill_rate
30
+ self.last_refill = time.monotonic() # Last refill time
31
+ self.log: List[Any] = []
32
+ self.turbo_mode = False
33
+
34
+ self.creation_time = time.monotonic()
35
+
36
+ self.num_requests = 0
37
+ self.num_released = 0
38
+ self.tokens_returned = 0
39
+
40
+ def turbo_mode_on(self):
41
+ """Set the refill rate to infinity."""
42
+ if self.turbo_mode:
43
+ pass
44
+ else:
45
+ # pass
46
+ self.turbo_mode = True
47
+ self.capacity = float("inf")
48
+ self.refill_rate = float("inf")
49
+
50
+ def turbo_mode_off(self):
51
+ """Restore the refill rate to its original value."""
52
+ self.turbo_mode = False
53
+ self.capacity = self._old_capacity
54
+ self.refill_rate = self._old_refill_rate
55
+
56
+ def __add__(self, other) -> "TokenBucket":
57
+ """Combine two token buckets.
58
+
59
+ The resulting bucket has the minimum capacity and refill rate of the two buckets.
60
+ This is useful, for example, if we have two calls to the same model on the same service but have different temperatures.
61
+ """
62
+ return TokenBucket(
63
+ bucket_name=self.bucket_name,
64
+ bucket_type=self.bucket_type,
65
+ capacity=min(self.capacity, other.capacity),
66
+ refill_rate=min(self.refill_rate, other.refill_rate),
67
+ )
68
+
69
+ def __repr__(self):
70
+ return f"TokenBucket(bucket_name={self.bucket_name}, bucket_type='{self.bucket_type}', capacity={self.capacity}, refill_rate={self.refill_rate})"
71
+
72
+ def add_tokens(self, tokens: Union[int, float]) -> None:
73
+ """Add tokens to the bucket, up to the maximum capacity.
74
+
75
+ :param tokens: The number of tokens to add to the bucket.
76
+
77
+ >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
78
+ >>> bucket.tokens
79
+ 10
80
+ >>> bucket.add_tokens(5)
81
+ >>> bucket.tokens
82
+ 10
83
+ """
84
+ self.tokens_returned += tokens
85
+ self.tokens = min(self.capacity, self.tokens + tokens)
86
+ self.log.append((time.monotonic(), self.tokens))
87
+
88
+ def refill(self) -> None:
89
+ """Refill the bucket with new tokens based on elapsed time.
90
+
91
+
92
+
93
+ >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
94
+ >>> bucket.tokens = 0
95
+ >>> bucket.refill()
96
+ >>> bucket.tokens > 0
97
+ True
98
+ """
99
+ """Refill the bucket with new tokens based on elapsed time."""
100
+ now = time.monotonic()
101
+ # print(f"Time is now: {now}; Last refill time: {self.last_refill}")
102
+ elapsed = now - self.last_refill
103
+ # print("Elapsed time: ", elapsed)
104
+ refill_amount = elapsed * self.refill_rate
105
+ self.tokens = min(self.capacity, self.tokens + refill_amount)
106
+ self.last_refill = now
107
+
108
+ if self.tokens < self.capacity:
109
+ pass
110
+ # print(f"Refilled. Current tokens: {self.tokens:.4f}")
111
+ # print(f"Elapsed time: {elapsed:.4f} seconds")
112
+ # print(f"Refill amount: {refill_amount:.4f}")
113
+
114
+ self.log.append((now, self.tokens))
115
+
116
+ def wait_time(self, requested_tokens: Union[float, int]) -> float:
117
+ """Calculate the time to wait for the requested number of tokens."""
118
+ # self.refill() # Update the current token count
119
+ if self.tokens >= requested_tokens:
120
+ return 0
121
+ return (requested_tokens - self.tokens) / self.refill_rate
122
+
123
+ async def get_tokens(
124
+ self, amount: Union[int, float] = 1, cheat_bucket_capacity=True
125
+ ) -> None:
126
+ """Wait for the specified number of tokens to become available.
127
+
128
+
129
+ :param amount: The number of tokens
130
+ :param warn: If True, warn if the requested amount exceeds the bucket capacity.
131
+
132
+ >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
133
+ >>> asyncio.run(bucket.get_tokens(5))
134
+ >>> bucket.tokens
135
+ 5
136
+ >>> asyncio.run(bucket.get_tokens(9))
137
+ >>> bucket.tokens < 1
138
+ True
139
+
140
+ >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=10, refill_rate=1)
141
+ >>> asyncio.run(bucket.get_tokens(11, cheat_bucket_capacity=False))
142
+ Traceback (most recent call last):
143
+ ...
144
+ ValueError: Requested amount exceeds bucket capacity. Bucket capacity: 10, requested amount: 11. As the bucket never overflows, the requested amount will never be available.
145
+ >>> asyncio.run(bucket.get_tokens(11, cheat_bucket_capacity=True))
146
+ >>> bucket.capacity
147
+ 12.100000000000001
148
+ """
149
+ self.num_requests += amount
150
+ if amount >= self.capacity:
151
+ if not cheat_bucket_capacity:
152
+ msg = f"Requested amount exceeds bucket capacity. Bucket capacity: {self.capacity}, requested amount: {amount}. As the bucket never overflows, the requested amount will never be available."
153
+ raise ValueError(msg)
154
+ else:
155
+ self.capacity = amount * 1.10
156
+ self._old_capacity = self.capacity
157
+
158
+ start_time = time.monotonic()
159
+ while True:
160
+ self.refill() # Refill based on elapsed time
161
+ if self.tokens >= amount:
162
+ self.tokens -= amount
163
+ break
164
+
165
+ wait_time = self.wait_time(amount)
166
+ if wait_time > 0:
167
+ await asyncio.sleep(wait_time)
168
+
169
+ self.num_released += amount
170
+ now = time.monotonic()
171
+ self.log.append((now, self.tokens))
172
+ return None
173
+
174
+ def get_log(self) -> list[tuple]:
175
+ return self.log
176
+
177
+ def visualize(self):
178
+ """Visualize the token bucket over time."""
179
+ times, tokens = zip(*self.get_log())
180
+ start_time = times[0]
181
+ times = [t - start_time for t in times] # Normalize time to start from 0
182
+ from matplotlib import pyplot as plt
183
+
184
+ plt.figure(figsize=(10, 6))
185
+ plt.plot(times, tokens, label="Tokens Available")
186
+ plt.xlabel("Time (seconds)", fontsize=12)
187
+ plt.ylabel("Number of Tokens", fontsize=12)
188
+ details = f"{self.bucket_name} ({self.bucket_type}) Bucket Usage Over Time\nCapacity: {self.capacity:.1f}, Refill Rate: {self.refill_rate:.1f}/second"
189
+ plt.title(details, fontsize=14)
190
+
191
+ plt.legend()
192
+ plt.grid(True)
193
+ plt.tight_layout()
194
+ plt.show()
195
+
196
+ def get_throughput(self, time_window: Optional[float] = None) -> float:
197
+ """
198
+ Calculate the empirical bucket throughput in tokens per minute for the specified time window.
199
+
200
+ :param time_window: The time window in seconds to calculate the throughput for.
201
+ :return: The throughput in tokens per minute.
202
+
203
+ >>> bucket = TokenBucket(bucket_name="test", bucket_type="test", capacity=100, refill_rate=10)
204
+ >>> asyncio.run(bucket.get_tokens(50))
205
+ >>> time.sleep(1) # Wait for 1 second
206
+ >>> asyncio.run(bucket.get_tokens(30))
207
+ >>> throughput = bucket.get_throughput(1)
208
+ >>> 4750 < throughput < 4850
209
+ True
210
+ """
211
+ now = time.monotonic()
212
+
213
+ if time_window is None:
214
+ start_time = self.creation_time
215
+ else:
216
+ start_time = now - time_window
217
+
218
+ if start_time < self.creation_time:
219
+ start_time = self.creation_time
220
+
221
+ elapsed_time = now - start_time
222
+
223
+ if elapsed_time == 0:
224
+ return self.num_released / 0.001
225
+
226
+ return (self.num_released / elapsed_time) * 60
227
+
228
+ # # Filter log entries within the time window
229
+ # relevant_log = [(t, tokens) for t, tokens in self.log if t >= start_time]
230
+
231
+ # if len(relevant_log) < 2:
232
+ # return 0 # Not enough data points to calculate throughput
233
+
234
+ # # Calculate total tokens used
235
+ # initial_tokens = relevant_log[0][1]
236
+ # final_tokens = relevant_log[-1][1]
237
+ # tokens_used = self.num_released - (final_tokens - initial_tokens)
238
+
239
+ # # Calculate actual time elapsed
240
+ # actual_time_elapsed = relevant_log[-1][0] - relevant_log[0][0]
241
+
242
+ # # Calculate throughput in tokens per minute
243
+ # throughput = (tokens_used / actual_time_elapsed) * 60
244
+
245
+ # return throughput
246
+
247
+
248
+ if __name__ == "__main__":
249
+ import doctest
250
+
251
+ doctest.testmod(optionflags=doctest.ELLIPSIS)