edsl 0.1.39.dev3__py3-none-any.whl → 0.1.39.dev4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- edsl/Base.py +413 -332
- edsl/BaseDiff.py +260 -260
- edsl/TemplateLoader.py +24 -24
- edsl/__init__.py +57 -49
- edsl/__version__.py +1 -1
- edsl/agents/Agent.py +1071 -867
- edsl/agents/AgentList.py +551 -413
- edsl/agents/Invigilator.py +284 -233
- edsl/agents/InvigilatorBase.py +257 -270
- edsl/agents/PromptConstructor.py +272 -354
- edsl/agents/QuestionInstructionPromptBuilder.py +128 -0
- edsl/agents/QuestionTemplateReplacementsBuilder.py +137 -0
- edsl/agents/__init__.py +2 -3
- edsl/agents/descriptors.py +99 -99
- edsl/agents/prompt_helpers.py +129 -129
- edsl/agents/question_option_processor.py +172 -0
- edsl/auto/AutoStudy.py +130 -117
- edsl/auto/StageBase.py +243 -230
- edsl/auto/StageGenerateSurvey.py +178 -178
- edsl/auto/StageLabelQuestions.py +125 -125
- edsl/auto/StagePersona.py +61 -61
- edsl/auto/StagePersonaDimensionValueRanges.py +88 -88
- edsl/auto/StagePersonaDimensionValues.py +74 -74
- edsl/auto/StagePersonaDimensions.py +69 -69
- edsl/auto/StageQuestions.py +74 -73
- edsl/auto/SurveyCreatorPipeline.py +21 -21
- edsl/auto/utilities.py +218 -224
- edsl/base/Base.py +279 -279
- edsl/config.py +177 -157
- edsl/conversation/Conversation.py +290 -290
- edsl/conversation/car_buying.py +59 -58
- edsl/conversation/chips.py +95 -95
- edsl/conversation/mug_negotiation.py +81 -81
- edsl/conversation/next_speaker_utilities.py +93 -93
- edsl/coop/CoopFunctionsMixin.py +15 -0
- edsl/coop/ExpectedParrotKeyHandler.py +125 -0
- edsl/coop/PriceFetcher.py +54 -54
- edsl/coop/__init__.py +2 -2
- edsl/coop/coop.py +1106 -1028
- edsl/coop/utils.py +131 -131
- edsl/data/Cache.py +573 -555
- edsl/data/CacheEntry.py +230 -233
- edsl/data/CacheHandler.py +168 -149
- edsl/data/RemoteCacheSync.py +186 -78
- edsl/data/SQLiteDict.py +292 -292
- edsl/data/__init__.py +5 -4
- edsl/data/hack.py +10 -0
- edsl/data/orm.py +10 -10
- edsl/data_transfer_models.py +74 -73
- edsl/enums.py +202 -175
- edsl/exceptions/BaseException.py +21 -21
- edsl/exceptions/__init__.py +54 -54
- edsl/exceptions/agents.py +54 -42
- edsl/exceptions/cache.py +5 -5
- edsl/exceptions/configuration.py +16 -16
- edsl/exceptions/coop.py +10 -10
- edsl/exceptions/data.py +14 -14
- edsl/exceptions/general.py +34 -34
- edsl/exceptions/inference_services.py +5 -0
- edsl/exceptions/jobs.py +33 -33
- edsl/exceptions/language_models.py +63 -63
- edsl/exceptions/prompts.py +15 -15
- edsl/exceptions/questions.py +109 -91
- edsl/exceptions/results.py +29 -29
- edsl/exceptions/scenarios.py +29 -22
- edsl/exceptions/surveys.py +37 -37
- edsl/inference_services/AnthropicService.py +106 -87
- edsl/inference_services/AvailableModelCacheHandler.py +184 -0
- edsl/inference_services/AvailableModelFetcher.py +215 -0
- edsl/inference_services/AwsBedrock.py +118 -120
- edsl/inference_services/AzureAI.py +215 -217
- edsl/inference_services/DeepInfraService.py +18 -18
- edsl/inference_services/GoogleService.py +143 -148
- edsl/inference_services/GroqService.py +20 -20
- edsl/inference_services/InferenceServiceABC.py +80 -147
- edsl/inference_services/InferenceServicesCollection.py +138 -97
- edsl/inference_services/MistralAIService.py +120 -123
- edsl/inference_services/OllamaService.py +18 -18
- edsl/inference_services/OpenAIService.py +236 -224
- edsl/inference_services/PerplexityService.py +160 -163
- edsl/inference_services/ServiceAvailability.py +135 -0
- edsl/inference_services/TestService.py +90 -89
- edsl/inference_services/TogetherAIService.py +172 -170
- edsl/inference_services/data_structures.py +134 -0
- edsl/inference_services/models_available_cache.py +118 -118
- edsl/inference_services/rate_limits_cache.py +25 -25
- edsl/inference_services/registry.py +41 -41
- edsl/inference_services/write_available.py +10 -10
- edsl/jobs/AnswerQuestionFunctionConstructor.py +223 -0
- edsl/jobs/Answers.py +43 -56
- edsl/jobs/FetchInvigilator.py +47 -0
- edsl/jobs/InterviewTaskManager.py +98 -0
- edsl/jobs/InterviewsConstructor.py +50 -0
- edsl/jobs/Jobs.py +823 -898
- edsl/jobs/JobsChecks.py +172 -147
- edsl/jobs/JobsComponentConstructor.py +189 -0
- edsl/jobs/JobsPrompts.py +270 -268
- edsl/jobs/JobsRemoteInferenceHandler.py +311 -239
- edsl/jobs/JobsRemoteInferenceLogger.py +239 -0
- edsl/jobs/RequestTokenEstimator.py +30 -0
- edsl/jobs/__init__.py +1 -1
- edsl/jobs/async_interview_runner.py +138 -0
- edsl/jobs/buckets/BucketCollection.py +104 -63
- edsl/jobs/buckets/ModelBuckets.py +65 -65
- edsl/jobs/buckets/TokenBucket.py +283 -251
- edsl/jobs/buckets/TokenBucketAPI.py +211 -0
- edsl/jobs/buckets/TokenBucketClient.py +191 -0
- edsl/jobs/check_survey_scenario_compatibility.py +85 -0
- edsl/jobs/data_structures.py +120 -0
- edsl/jobs/decorators.py +35 -0
- edsl/jobs/interviews/Interview.py +396 -661
- edsl/jobs/interviews/InterviewExceptionCollection.py +99 -99
- edsl/jobs/interviews/InterviewExceptionEntry.py +186 -186
- edsl/jobs/interviews/InterviewStatistic.py +63 -63
- edsl/jobs/interviews/InterviewStatisticsCollection.py +25 -25
- edsl/jobs/interviews/InterviewStatusDictionary.py +78 -78
- edsl/jobs/interviews/InterviewStatusLog.py +92 -92
- edsl/jobs/interviews/ReportErrors.py +66 -66
- edsl/jobs/interviews/interview_status_enum.py +9 -9
- edsl/jobs/jobs_status_enums.py +9 -0
- edsl/jobs/loggers/HTMLTableJobLogger.py +304 -0
- edsl/jobs/results_exceptions_handler.py +98 -0
- edsl/jobs/runners/JobsRunnerAsyncio.py +151 -466
- edsl/jobs/runners/JobsRunnerStatus.py +297 -330
- edsl/jobs/tasks/QuestionTaskCreator.py +244 -242
- edsl/jobs/tasks/TaskCreators.py +64 -64
- edsl/jobs/tasks/TaskHistory.py +470 -450
- edsl/jobs/tasks/TaskStatusLog.py +23 -23
- edsl/jobs/tasks/task_status_enum.py +161 -163
- edsl/jobs/tokens/InterviewTokenUsage.py +27 -27
- edsl/jobs/tokens/TokenUsage.py +34 -34
- edsl/language_models/ComputeCost.py +63 -0
- edsl/language_models/LanguageModel.py +626 -668
- edsl/language_models/ModelList.py +164 -155
- edsl/language_models/PriceManager.py +127 -0
- edsl/language_models/RawResponseHandler.py +106 -0
- edsl/language_models/RegisterLanguageModelsMeta.py +184 -184
- edsl/language_models/ServiceDataSources.py +0 -0
- edsl/language_models/__init__.py +2 -3
- edsl/language_models/fake_openai_call.py +15 -15
- edsl/language_models/fake_openai_service.py +61 -61
- edsl/language_models/key_management/KeyLookup.py +63 -0
- edsl/language_models/key_management/KeyLookupBuilder.py +273 -0
- edsl/language_models/key_management/KeyLookupCollection.py +38 -0
- edsl/language_models/key_management/__init__.py +0 -0
- edsl/language_models/key_management/models.py +131 -0
- edsl/language_models/model.py +256 -0
- edsl/language_models/repair.py +156 -156
- edsl/language_models/utilities.py +65 -64
- edsl/notebooks/Notebook.py +263 -258
- edsl/notebooks/NotebookToLaTeX.py +142 -0
- edsl/notebooks/__init__.py +1 -1
- edsl/prompts/Prompt.py +352 -362
- edsl/prompts/__init__.py +2 -2
- edsl/questions/ExceptionExplainer.py +77 -0
- edsl/questions/HTMLQuestion.py +103 -0
- edsl/questions/QuestionBase.py +518 -664
- edsl/questions/QuestionBasePromptsMixin.py +221 -217
- edsl/questions/QuestionBudget.py +227 -227
- edsl/questions/QuestionCheckBox.py +359 -359
- edsl/questions/QuestionExtract.py +180 -182
- edsl/questions/QuestionFreeText.py +113 -114
- edsl/questions/QuestionFunctional.py +166 -166
- edsl/questions/QuestionList.py +223 -231
- edsl/questions/QuestionMatrix.py +265 -0
- edsl/questions/QuestionMultipleChoice.py +330 -286
- edsl/questions/QuestionNumerical.py +151 -153
- edsl/questions/QuestionRank.py +314 -324
- edsl/questions/Quick.py +41 -41
- edsl/questions/SimpleAskMixin.py +74 -73
- edsl/questions/__init__.py +27 -26
- edsl/questions/{AnswerValidatorMixin.py → answer_validator_mixin.py} +334 -289
- edsl/questions/compose_questions.py +98 -98
- edsl/questions/data_structures.py +20 -0
- edsl/questions/decorators.py +21 -21
- edsl/questions/derived/QuestionLikertFive.py +76 -76
- edsl/questions/derived/QuestionLinearScale.py +90 -87
- edsl/questions/derived/QuestionTopK.py +93 -93
- edsl/questions/derived/QuestionYesNo.py +82 -82
- edsl/questions/descriptors.py +427 -413
- edsl/questions/loop_processor.py +149 -0
- edsl/questions/prompt_templates/question_budget.jinja +13 -13
- edsl/questions/prompt_templates/question_checkbox.jinja +32 -32
- edsl/questions/prompt_templates/question_extract.jinja +11 -11
- edsl/questions/prompt_templates/question_free_text.jinja +3 -3
- edsl/questions/prompt_templates/question_linear_scale.jinja +11 -11
- edsl/questions/prompt_templates/question_list.jinja +17 -17
- edsl/questions/prompt_templates/question_multiple_choice.jinja +33 -33
- edsl/questions/prompt_templates/question_numerical.jinja +36 -36
- edsl/questions/{QuestionBaseGenMixin.py → question_base_gen_mixin.py} +168 -161
- edsl/questions/question_registry.py +177 -177
- edsl/questions/{RegisterQuestionsMeta.py → register_questions_meta.py} +71 -71
- edsl/questions/{ResponseValidatorABC.py → response_validator_abc.py} +188 -174
- edsl/questions/response_validator_factory.py +34 -0
- edsl/questions/settings.py +12 -12
- edsl/questions/templates/budget/answering_instructions.jinja +7 -7
- edsl/questions/templates/budget/question_presentation.jinja +7 -7
- edsl/questions/templates/checkbox/answering_instructions.jinja +10 -10
- edsl/questions/templates/checkbox/question_presentation.jinja +22 -22
- edsl/questions/templates/extract/answering_instructions.jinja +7 -7
- edsl/questions/templates/likert_five/answering_instructions.jinja +10 -10
- edsl/questions/templates/likert_five/question_presentation.jinja +11 -11
- edsl/questions/templates/linear_scale/answering_instructions.jinja +5 -5
- edsl/questions/templates/linear_scale/question_presentation.jinja +5 -5
- edsl/questions/templates/list/answering_instructions.jinja +3 -3
- edsl/questions/templates/list/question_presentation.jinja +5 -5
- edsl/questions/templates/matrix/__init__.py +1 -0
- edsl/questions/templates/matrix/answering_instructions.jinja +5 -0
- edsl/questions/templates/matrix/question_presentation.jinja +20 -0
- edsl/questions/templates/multiple_choice/answering_instructions.jinja +9 -9
- edsl/questions/templates/multiple_choice/question_presentation.jinja +11 -11
- edsl/questions/templates/numerical/answering_instructions.jinja +6 -6
- edsl/questions/templates/numerical/question_presentation.jinja +6 -6
- edsl/questions/templates/rank/answering_instructions.jinja +11 -11
- edsl/questions/templates/rank/question_presentation.jinja +15 -15
- edsl/questions/templates/top_k/answering_instructions.jinja +8 -8
- edsl/questions/templates/top_k/question_presentation.jinja +22 -22
- edsl/questions/templates/yes_no/answering_instructions.jinja +6 -6
- edsl/questions/templates/yes_no/question_presentation.jinja +11 -11
- edsl/results/CSSParameterizer.py +108 -108
- edsl/results/Dataset.py +587 -424
- edsl/results/DatasetExportMixin.py +594 -731
- edsl/results/DatasetTree.py +295 -275
- edsl/results/MarkdownToDocx.py +122 -0
- edsl/results/MarkdownToPDF.py +111 -0
- edsl/results/Result.py +557 -465
- edsl/results/Results.py +1183 -1165
- edsl/results/ResultsExportMixin.py +45 -43
- edsl/results/ResultsGGMixin.py +121 -121
- edsl/results/TableDisplay.py +125 -198
- edsl/results/TextEditor.py +50 -0
- edsl/results/__init__.py +2 -2
- edsl/results/file_exports.py +252 -0
- edsl/results/{ResultsFetchMixin.py → results_fetch_mixin.py} +33 -33
- edsl/results/{Selector.py → results_selector.py} +145 -135
- edsl/results/{ResultsToolsMixin.py → results_tools_mixin.py} +98 -98
- edsl/results/smart_objects.py +96 -0
- edsl/results/table_data_class.py +12 -0
- edsl/results/table_display.css +77 -77
- edsl/results/table_renderers.py +118 -0
- edsl/results/tree_explore.py +115 -115
- edsl/scenarios/ConstructDownloadLink.py +109 -0
- edsl/scenarios/DocumentChunker.py +102 -0
- edsl/scenarios/DocxScenario.py +16 -0
- edsl/scenarios/FileStore.py +511 -632
- edsl/scenarios/PdfExtractor.py +40 -0
- edsl/scenarios/Scenario.py +498 -601
- edsl/scenarios/ScenarioHtmlMixin.py +65 -64
- edsl/scenarios/ScenarioList.py +1458 -1287
- edsl/scenarios/ScenarioListExportMixin.py +45 -52
- edsl/scenarios/ScenarioListPdfMixin.py +239 -261
- edsl/scenarios/__init__.py +3 -4
- edsl/scenarios/directory_scanner.py +96 -0
- edsl/scenarios/file_methods.py +85 -0
- edsl/scenarios/handlers/__init__.py +13 -0
- edsl/scenarios/handlers/csv.py +38 -0
- edsl/scenarios/handlers/docx.py +76 -0
- edsl/scenarios/handlers/html.py +37 -0
- edsl/scenarios/handlers/json.py +111 -0
- edsl/scenarios/handlers/latex.py +5 -0
- edsl/scenarios/handlers/md.py +51 -0
- edsl/scenarios/handlers/pdf.py +68 -0
- edsl/scenarios/handlers/png.py +39 -0
- edsl/scenarios/handlers/pptx.py +105 -0
- edsl/scenarios/handlers/py.py +294 -0
- edsl/scenarios/handlers/sql.py +313 -0
- edsl/scenarios/handlers/sqlite.py +149 -0
- edsl/scenarios/handlers/txt.py +33 -0
- edsl/scenarios/{ScenarioJoin.py → scenario_join.py} +131 -127
- edsl/scenarios/scenario_selector.py +156 -0
- edsl/shared.py +1 -1
- edsl/study/ObjectEntry.py +173 -173
- edsl/study/ProofOfWork.py +113 -113
- edsl/study/SnapShot.py +80 -80
- edsl/study/Study.py +521 -528
- edsl/study/__init__.py +4 -4
- edsl/surveys/ConstructDAG.py +92 -0
- edsl/surveys/DAG.py +148 -148
- edsl/surveys/EditSurvey.py +221 -0
- edsl/surveys/InstructionHandler.py +100 -0
- edsl/surveys/Memory.py +31 -31
- edsl/surveys/MemoryManagement.py +72 -0
- edsl/surveys/MemoryPlan.py +244 -244
- edsl/surveys/Rule.py +327 -326
- edsl/surveys/RuleCollection.py +385 -387
- edsl/surveys/RuleManager.py +172 -0
- edsl/surveys/Simulator.py +75 -0
- edsl/surveys/Survey.py +1280 -1801
- edsl/surveys/SurveyCSS.py +273 -261
- edsl/surveys/SurveyExportMixin.py +259 -259
- edsl/surveys/{SurveyFlowVisualizationMixin.py → SurveyFlowVisualization.py} +181 -179
- edsl/surveys/SurveyQualtricsImport.py +284 -284
- edsl/surveys/SurveyToApp.py +141 -0
- edsl/surveys/__init__.py +5 -3
- edsl/surveys/base.py +53 -53
- edsl/surveys/descriptors.py +60 -56
- edsl/surveys/instructions/ChangeInstruction.py +48 -49
- edsl/surveys/instructions/Instruction.py +56 -65
- edsl/surveys/instructions/InstructionCollection.py +82 -77
- edsl/templates/error_reporting/base.html +23 -23
- edsl/templates/error_reporting/exceptions_by_model.html +34 -34
- edsl/templates/error_reporting/exceptions_by_question_name.html +16 -16
- edsl/templates/error_reporting/exceptions_by_type.html +16 -16
- edsl/templates/error_reporting/interview_details.html +115 -115
- edsl/templates/error_reporting/interviews.html +19 -19
- edsl/templates/error_reporting/overview.html +4 -4
- edsl/templates/error_reporting/performance_plot.html +1 -1
- edsl/templates/error_reporting/report.css +73 -73
- edsl/templates/error_reporting/report.html +117 -117
- edsl/templates/error_reporting/report.js +25 -25
- edsl/test_h +1 -0
- edsl/tools/__init__.py +1 -1
- edsl/tools/clusters.py +192 -192
- edsl/tools/embeddings.py +27 -27
- edsl/tools/embeddings_plotting.py +118 -118
- edsl/tools/plotting.py +112 -112
- edsl/tools/summarize.py +18 -18
- edsl/utilities/PrettyList.py +56 -0
- edsl/utilities/SystemInfo.py +28 -28
- edsl/utilities/__init__.py +22 -22
- edsl/utilities/ast_utilities.py +25 -25
- edsl/utilities/data/Registry.py +6 -6
- edsl/utilities/data/__init__.py +1 -1
- edsl/utilities/data/scooter_results.json +1 -1
- edsl/utilities/decorators.py +77 -77
- edsl/utilities/gcp_bucket/cloud_storage.py +96 -96
- edsl/utilities/gcp_bucket/example.py +50 -0
- edsl/utilities/interface.py +627 -627
- edsl/utilities/is_notebook.py +18 -0
- edsl/utilities/is_valid_variable_name.py +11 -0
- edsl/utilities/naming_utilities.py +263 -263
- edsl/utilities/remove_edsl_version.py +24 -0
- edsl/utilities/repair_functions.py +28 -28
- edsl/utilities/restricted_python.py +70 -70
- edsl/utilities/utilities.py +436 -424
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/LICENSE +21 -21
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/METADATA +13 -11
- edsl-0.1.39.dev4.dist-info/RECORD +361 -0
- edsl/language_models/KeyLookup.py +0 -30
- edsl/language_models/registry.py +0 -190
- edsl/language_models/unused/ReplicateBase.py +0 -83
- edsl/results/ResultsDBMixin.py +0 -238
- edsl-0.1.39.dev3.dist-info/RECORD +0 -277
- {edsl-0.1.39.dev3.dist-info → edsl-0.1.39.dev4.dist-info}/WHEEL +0 -0
@@ -1,163 +1,160 @@
|
|
1
|
-
import aiohttp
|
2
|
-
import json
|
3
|
-
import requests
|
4
|
-
from typing import Any, List, Optional
|
5
|
-
from edsl.inference_services.rate_limits_cache import rate_limits
|
6
|
-
|
7
|
-
# from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
|
8
|
-
from edsl.language_models import LanguageModel
|
9
|
-
|
10
|
-
from edsl.inference_services.OpenAIService import OpenAIService
|
11
|
-
|
12
|
-
|
13
|
-
class PerplexityService(OpenAIService):
|
14
|
-
"""Perplexity service class."""
|
15
|
-
|
16
|
-
_inference_service_ = "perplexity"
|
17
|
-
_env_key_name_ = "PERPLEXITY_API_KEY"
|
18
|
-
_base_url_ = "https://api.perplexity.ai"
|
19
|
-
_models_list_cache: List[str] = []
|
20
|
-
# default perplexity parameters
|
21
|
-
_parameters_ = {
|
22
|
-
"temperature": 0.5,
|
23
|
-
"max_tokens": 1000,
|
24
|
-
"top_p": 1,
|
25
|
-
"logprobs": False,
|
26
|
-
"top_logprobs": 3,
|
27
|
-
}
|
28
|
-
|
29
|
-
@classmethod
|
30
|
-
def available(cls) -> List[str]:
|
31
|
-
return [
|
32
|
-
"llama-3.1-sonar-huge-128k-online",
|
33
|
-
"llama-3.1-sonar-large-128k-online",
|
34
|
-
"llama-3.1-sonar-small-128k-online",
|
35
|
-
]
|
36
|
-
|
37
|
-
@classmethod
|
38
|
-
def create_model(
|
39
|
-
cls, model_name="llama-3.1-sonar-large-128k-online", model_class_name=None
|
40
|
-
) -> LanguageModel:
|
41
|
-
if model_class_name is None:
|
42
|
-
model_class_name = cls.to_class_name(model_name)
|
43
|
-
|
44
|
-
class LLM(LanguageModel):
|
45
|
-
"""
|
46
|
-
Child class of LanguageModel for interacting with Perplexity models
|
47
|
-
"""
|
48
|
-
|
49
|
-
key_sequence = cls.key_sequence
|
50
|
-
usage_sequence = cls.usage_sequence
|
51
|
-
input_token_name = cls.input_token_name
|
52
|
-
output_token_name = cls.output_token_name
|
53
|
-
|
54
|
-
|
55
|
-
|
56
|
-
|
57
|
-
|
58
|
-
|
59
|
-
|
60
|
-
|
61
|
-
"
|
62
|
-
"
|
63
|
-
"
|
64
|
-
"
|
65
|
-
|
66
|
-
|
67
|
-
|
68
|
-
|
69
|
-
|
70
|
-
def
|
71
|
-
return cls.
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
def
|
78
|
-
|
79
|
-
|
80
|
-
|
81
|
-
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
|
86
|
-
|
87
|
-
|
88
|
-
|
89
|
-
|
90
|
-
|
91
|
-
|
92
|
-
|
93
|
-
|
94
|
-
|
95
|
-
|
96
|
-
headers =
|
97
|
-
|
98
|
-
|
99
|
-
|
100
|
-
|
101
|
-
|
102
|
-
|
103
|
-
|
104
|
-
|
105
|
-
|
106
|
-
|
107
|
-
|
108
|
-
|
109
|
-
|
110
|
-
|
111
|
-
|
112
|
-
|
113
|
-
|
114
|
-
|
115
|
-
|
116
|
-
|
117
|
-
|
118
|
-
|
119
|
-
|
120
|
-
|
121
|
-
|
122
|
-
|
123
|
-
|
124
|
-
|
125
|
-
|
126
|
-
|
127
|
-
|
128
|
-
|
129
|
-
|
130
|
-
|
131
|
-
|
132
|
-
|
133
|
-
|
134
|
-
|
135
|
-
|
136
|
-
|
137
|
-
|
138
|
-
|
139
|
-
|
140
|
-
|
141
|
-
|
142
|
-
|
143
|
-
|
144
|
-
|
145
|
-
"
|
146
|
-
"
|
147
|
-
"
|
148
|
-
"
|
149
|
-
"
|
150
|
-
"
|
151
|
-
|
152
|
-
|
153
|
-
|
154
|
-
|
155
|
-
|
156
|
-
|
157
|
-
|
158
|
-
|
159
|
-
|
160
|
-
|
161
|
-
LLM.__name__ = "LanguageModel"
|
162
|
-
|
163
|
-
return LLM
|
1
|
+
import aiohttp
|
2
|
+
import json
|
3
|
+
import requests
|
4
|
+
from typing import Any, List, Optional
|
5
|
+
from edsl.inference_services.rate_limits_cache import rate_limits
|
6
|
+
|
7
|
+
# from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
|
8
|
+
from edsl.language_models import LanguageModel
|
9
|
+
|
10
|
+
from edsl.inference_services.OpenAIService import OpenAIService
|
11
|
+
|
12
|
+
|
13
|
+
class PerplexityService(OpenAIService):
|
14
|
+
"""Perplexity service class."""
|
15
|
+
|
16
|
+
_inference_service_ = "perplexity"
|
17
|
+
_env_key_name_ = "PERPLEXITY_API_KEY"
|
18
|
+
_base_url_ = "https://api.perplexity.ai"
|
19
|
+
_models_list_cache: List[str] = []
|
20
|
+
# default perplexity parameters
|
21
|
+
_parameters_ = {
|
22
|
+
"temperature": 0.5,
|
23
|
+
"max_tokens": 1000,
|
24
|
+
"top_p": 1,
|
25
|
+
"logprobs": False,
|
26
|
+
"top_logprobs": 3,
|
27
|
+
}
|
28
|
+
|
29
|
+
@classmethod
|
30
|
+
def available(cls) -> List[str]:
|
31
|
+
return [
|
32
|
+
"llama-3.1-sonar-huge-128k-online",
|
33
|
+
"llama-3.1-sonar-large-128k-online",
|
34
|
+
"llama-3.1-sonar-small-128k-online",
|
35
|
+
]
|
36
|
+
|
37
|
+
@classmethod
|
38
|
+
def create_model(
|
39
|
+
cls, model_name="llama-3.1-sonar-large-128k-online", model_class_name=None
|
40
|
+
) -> LanguageModel:
|
41
|
+
if model_class_name is None:
|
42
|
+
model_class_name = cls.to_class_name(model_name)
|
43
|
+
|
44
|
+
class LLM(LanguageModel):
|
45
|
+
"""
|
46
|
+
Child class of LanguageModel for interacting with Perplexity models
|
47
|
+
"""
|
48
|
+
|
49
|
+
key_sequence = cls.key_sequence
|
50
|
+
usage_sequence = cls.usage_sequence
|
51
|
+
input_token_name = cls.input_token_name
|
52
|
+
output_token_name = cls.output_token_name
|
53
|
+
|
54
|
+
_inference_service_ = cls._inference_service_
|
55
|
+
_model_ = model_name
|
56
|
+
|
57
|
+
_parameters_ = {
|
58
|
+
"temperature": 0.5,
|
59
|
+
"max_tokens": 1000,
|
60
|
+
"top_p": 1,
|
61
|
+
"frequency_penalty": 1,
|
62
|
+
"presence_penalty": 0,
|
63
|
+
# "logprobs": False, # Enable this returns 'Neither or both of logprobs and top_logprobs must be set.
|
64
|
+
# "top_logprobs": 3,
|
65
|
+
}
|
66
|
+
|
67
|
+
def sync_client(self):
|
68
|
+
return cls.sync_client()
|
69
|
+
|
70
|
+
def async_client(self):
|
71
|
+
return cls.async_client()
|
72
|
+
|
73
|
+
@classmethod
|
74
|
+
def available(cls) -> list[str]:
|
75
|
+
return cls.sync_client().models.list()
|
76
|
+
|
77
|
+
def get_headers(self) -> dict[str, Any]:
|
78
|
+
client = self.sync_client()
|
79
|
+
response = client.chat.completions.with_raw_response.create(
|
80
|
+
messages=[
|
81
|
+
{
|
82
|
+
"role": "user",
|
83
|
+
"content": "Say this is a test",
|
84
|
+
}
|
85
|
+
],
|
86
|
+
model=self.model,
|
87
|
+
)
|
88
|
+
return dict(response.headers)
|
89
|
+
|
90
|
+
def get_rate_limits(self) -> dict[str, Any]:
|
91
|
+
try:
|
92
|
+
if "openai" in rate_limits:
|
93
|
+
headers = rate_limits["openai"]
|
94
|
+
|
95
|
+
else:
|
96
|
+
headers = self.get_headers()
|
97
|
+
|
98
|
+
except Exception as e:
|
99
|
+
return {
|
100
|
+
"rpm": 10_000,
|
101
|
+
"tpm": 2_000_000,
|
102
|
+
}
|
103
|
+
else:
|
104
|
+
return {
|
105
|
+
"rpm": int(headers["x-ratelimit-limit-requests"]),
|
106
|
+
"tpm": int(headers["x-ratelimit-limit-tokens"]),
|
107
|
+
}
|
108
|
+
|
109
|
+
async def async_execute_model_call(
|
110
|
+
self,
|
111
|
+
user_prompt: str,
|
112
|
+
system_prompt: str = "",
|
113
|
+
files_list: Optional[List["Files"]] = None,
|
114
|
+
invigilator: Optional[
|
115
|
+
"InvigilatorAI"
|
116
|
+
] = None, # TBD - can eventually be used for function-calling
|
117
|
+
) -> dict[str, Any]:
|
118
|
+
"""Calls the OpenAI API and returns the API response."""
|
119
|
+
if files_list:
|
120
|
+
encoded_image = files_list[0].base64_string
|
121
|
+
content = [{"type": "text", "text": user_prompt}]
|
122
|
+
content.append(
|
123
|
+
{
|
124
|
+
"type": "image_url",
|
125
|
+
"image_url": {
|
126
|
+
"url": f"data:image/jpeg;base64,{encoded_image}"
|
127
|
+
},
|
128
|
+
}
|
129
|
+
)
|
130
|
+
else:
|
131
|
+
content = user_prompt
|
132
|
+
client = self.async_client()
|
133
|
+
|
134
|
+
messages = [
|
135
|
+
{"role": "system", "content": system_prompt},
|
136
|
+
{"role": "user", "content": content},
|
137
|
+
]
|
138
|
+
if system_prompt == "" and self.omit_system_prompt_if_empty:
|
139
|
+
messages = messages[1:]
|
140
|
+
|
141
|
+
params = {
|
142
|
+
"model": self.model,
|
143
|
+
"messages": messages,
|
144
|
+
"temperature": self.temperature,
|
145
|
+
"max_tokens": self.max_tokens,
|
146
|
+
"top_p": self.top_p,
|
147
|
+
"frequency_penalty": self.frequency_penalty,
|
148
|
+
"presence_penalty": self.presence_penalty,
|
149
|
+
# "logprobs": self.logprobs,
|
150
|
+
# "top_logprobs": self.top_logprobs if self.logprobs else None,
|
151
|
+
}
|
152
|
+
try:
|
153
|
+
response = await client.chat.completions.create(**params)
|
154
|
+
except Exception as e:
|
155
|
+
print(e, flush=True)
|
156
|
+
return response.model_dump()
|
157
|
+
|
158
|
+
LLM.__name__ = "LanguageModel"
|
159
|
+
|
160
|
+
return LLM
|
@@ -0,0 +1,135 @@
|
|
1
|
+
from enum import Enum
|
2
|
+
from typing import List, Optional, TYPE_CHECKING
|
3
|
+
from functools import partial
|
4
|
+
import warnings
|
5
|
+
|
6
|
+
from edsl.inference_services.data_structures import AvailableModels, ModelNamesList
|
7
|
+
|
8
|
+
if TYPE_CHECKING:
|
9
|
+
from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
|
10
|
+
|
11
|
+
|
12
|
+
class ModelSource(Enum):
|
13
|
+
LOCAL = "local"
|
14
|
+
COOP = "coop"
|
15
|
+
CACHE = "cache"
|
16
|
+
|
17
|
+
|
18
|
+
class ServiceAvailability:
|
19
|
+
"""This class is responsible for fetching the available models from different sources."""
|
20
|
+
|
21
|
+
_coop_model_list = None
|
22
|
+
|
23
|
+
def __init__(self, source_order: Optional[List[ModelSource]] = None):
|
24
|
+
"""
|
25
|
+
Initialize with custom source order.
|
26
|
+
Default order is LOCAL -> COOP -> CACHE
|
27
|
+
"""
|
28
|
+
self.source_order = source_order or [
|
29
|
+
ModelSource.LOCAL,
|
30
|
+
ModelSource.COOP,
|
31
|
+
ModelSource.CACHE,
|
32
|
+
]
|
33
|
+
|
34
|
+
# Map sources to their fetch functions
|
35
|
+
self._source_fetchers = {
|
36
|
+
ModelSource.LOCAL: self._fetch_from_local_service,
|
37
|
+
ModelSource.COOP: self._fetch_from_coop,
|
38
|
+
ModelSource.CACHE: self._fetch_from_cache,
|
39
|
+
}
|
40
|
+
|
41
|
+
@classmethod
|
42
|
+
def models_from_coop(cls) -> AvailableModels:
|
43
|
+
if not cls._coop_model_list:
|
44
|
+
from edsl.coop.coop import Coop
|
45
|
+
|
46
|
+
c = Coop()
|
47
|
+
coop_model_list = c.fetch_models()
|
48
|
+
cls._coop_model_list = coop_model_list
|
49
|
+
return cls._coop_model_list
|
50
|
+
|
51
|
+
def get_service_available(
|
52
|
+
self, service: "InferenceServiceABC", warn: bool = False
|
53
|
+
) -> ModelNamesList:
|
54
|
+
"""
|
55
|
+
Try to fetch available models from sources in specified order.
|
56
|
+
Returns first successful result.
|
57
|
+
"""
|
58
|
+
last_error = None
|
59
|
+
|
60
|
+
for source in self.source_order:
|
61
|
+
try:
|
62
|
+
fetch_func = partial(self._source_fetchers[source], service)
|
63
|
+
result = fetch_func()
|
64
|
+
|
65
|
+
# Cache successful result
|
66
|
+
service._models_list_cache = result
|
67
|
+
return result
|
68
|
+
|
69
|
+
except Exception as e:
|
70
|
+
last_error = e
|
71
|
+
if warn:
|
72
|
+
self._warn_source_failed(service, source)
|
73
|
+
continue
|
74
|
+
|
75
|
+
# If we get here, all sources failed
|
76
|
+
raise RuntimeError(
|
77
|
+
f"All sources failed to fetch models. Last error: {last_error}"
|
78
|
+
)
|
79
|
+
|
80
|
+
@staticmethod
|
81
|
+
def _fetch_from_local_service(service: "InferenceServiceABC") -> ModelNamesList:
|
82
|
+
"""Attempt to fetch models directly from the service."""
|
83
|
+
return service.available()
|
84
|
+
|
85
|
+
@classmethod
|
86
|
+
def _fetch_from_coop(cls, service: "InferenceServiceABC") -> ModelNamesList:
|
87
|
+
"""Fetch models from Coop."""
|
88
|
+
models_from_coop = cls.models_from_coop()
|
89
|
+
return models_from_coop.get(service._inference_service_, [])
|
90
|
+
|
91
|
+
@staticmethod
|
92
|
+
def _fetch_from_cache(service: "InferenceServiceABC") -> ModelNamesList:
|
93
|
+
"""Fetch models from local cache."""
|
94
|
+
from edsl.inference_services.models_available_cache import models_available
|
95
|
+
|
96
|
+
return models_available.get(service._inference_service_, [])
|
97
|
+
|
98
|
+
def _warn_source_failed(self, service: "InferenceServiceABC", source: ModelSource):
|
99
|
+
"""Display appropriate warning message based on failed source."""
|
100
|
+
messages = {
|
101
|
+
ModelSource.LOCAL: f"""Error getting models for {service._inference_service_}.
|
102
|
+
Check that you have properly stored your Expected Parrot API key and activated remote inference,
|
103
|
+
or stored your own API keys for the language models that you want to use.
|
104
|
+
See https://docs.expectedparrot.com/en/latest/api_keys.html for instructions on storing API keys.
|
105
|
+
Trying next source.""",
|
106
|
+
ModelSource.COOP: f"Error getting models from Coop for {service._inference_service_}. Trying next source.",
|
107
|
+
ModelSource.CACHE: f"Error getting models from cache for {service._inference_service_}.",
|
108
|
+
}
|
109
|
+
warnings.warn(messages[source], UserWarning)
|
110
|
+
|
111
|
+
|
112
|
+
if __name__ == "__main__":
|
113
|
+
# sa = ServiceAvailability()
|
114
|
+
# models_from_coop = sa.models_from_coop()
|
115
|
+
# print(models_from_coop)
|
116
|
+
from edsl.inference_services.OpenAIService import OpenAIService
|
117
|
+
|
118
|
+
openai_models = ServiceAvailability._fetch_from_local_service(OpenAIService())
|
119
|
+
print(openai_models)
|
120
|
+
|
121
|
+
# Example usage:
|
122
|
+
"""
|
123
|
+
# Default order (LOCAL -> COOP -> CACHE)
|
124
|
+
availability = ServiceAvailability()
|
125
|
+
|
126
|
+
# Custom order (COOP -> LOCAL -> CACHE)
|
127
|
+
availability_coop_first = ServiceAvailability([
|
128
|
+
ModelSource.COOP,
|
129
|
+
ModelSource.LOCAL,
|
130
|
+
ModelSource.CACHE
|
131
|
+
])
|
132
|
+
|
133
|
+
# Get available models using custom order
|
134
|
+
models = availability_coop_first.get_service_available(service, warn=True)
|
135
|
+
"""
|
@@ -1,89 +1,90 @@
|
|
1
|
-
from typing import Any, List, Optional
|
2
|
-
import os
|
3
|
-
import asyncio
|
4
|
-
from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
|
5
|
-
from edsl.language_models import LanguageModel
|
6
|
-
from edsl.inference_services.rate_limits_cache import rate_limits
|
7
|
-
from edsl.utilities.utilities import fix_partial_correct_response
|
8
|
-
|
9
|
-
from edsl.enums import InferenceServiceType
|
10
|
-
import random
|
11
|
-
|
12
|
-
|
13
|
-
class TestService(InferenceServiceABC):
|
14
|
-
"""OpenAI service class."""
|
15
|
-
|
16
|
-
_inference_service_ = "test"
|
17
|
-
_env_key_name_ = None
|
18
|
-
_base_url_ = None
|
19
|
-
|
20
|
-
_sync_client_ = None
|
21
|
-
_async_client_ = None
|
22
|
-
|
23
|
-
_sync_client_instance = None
|
24
|
-
_async_client_instance = None
|
25
|
-
|
26
|
-
key_sequence = None
|
27
|
-
usage_sequence = None
|
28
|
-
model_exclude_list = []
|
29
|
-
input_token_name = "prompt_tokens"
|
30
|
-
output_token_name = "completion_tokens"
|
31
|
-
|
32
|
-
@classmethod
|
33
|
-
def available(cls) -> list[str]:
|
34
|
-
return ["test"]
|
35
|
-
|
36
|
-
@classmethod
|
37
|
-
def create_model(cls, model_name, model_class_name=None) -> LanguageModel:
|
38
|
-
throw_exception = False
|
39
|
-
|
40
|
-
class TestServiceLanguageModel(LanguageModel):
|
41
|
-
_model_ = "test"
|
42
|
-
_parameters_ = {"temperature": 0.5}
|
43
|
-
_inference_service_ = InferenceServiceType.TEST.value
|
44
|
-
usage_sequence = ["usage"]
|
45
|
-
key_sequence = ["message", 0, "text"]
|
46
|
-
input_token_name = cls.input_token_name
|
47
|
-
output_token_name = cls.output_token_name
|
48
|
-
_rpm = 1000
|
49
|
-
_tpm = 100000
|
50
|
-
|
51
|
-
@property
|
52
|
-
def _canned_response(self):
|
53
|
-
if hasattr(self, "canned_response"):
|
54
|
-
|
55
|
-
|
56
|
-
|
57
|
-
|
58
|
-
|
59
|
-
|
60
|
-
|
61
|
-
|
62
|
-
|
63
|
-
|
64
|
-
|
65
|
-
|
66
|
-
|
67
|
-
|
68
|
-
if hasattr(self, "
|
69
|
-
|
70
|
-
|
71
|
-
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
|
78
|
-
|
79
|
-
|
80
|
-
|
81
|
-
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
|
86
|
-
"
|
87
|
-
|
88
|
-
|
89
|
-
|
1
|
+
from typing import Any, List, Optional
|
2
|
+
import os
|
3
|
+
import asyncio
|
4
|
+
from edsl.inference_services.InferenceServiceABC import InferenceServiceABC
|
5
|
+
from edsl.language_models.LanguageModel import LanguageModel
|
6
|
+
from edsl.inference_services.rate_limits_cache import rate_limits
|
7
|
+
from edsl.utilities.utilities import fix_partial_correct_response
|
8
|
+
|
9
|
+
from edsl.enums import InferenceServiceType
|
10
|
+
import random
|
11
|
+
|
12
|
+
|
13
|
+
class TestService(InferenceServiceABC):
|
14
|
+
"""OpenAI service class."""
|
15
|
+
|
16
|
+
_inference_service_ = "test"
|
17
|
+
_env_key_name_ = None
|
18
|
+
_base_url_ = None
|
19
|
+
|
20
|
+
_sync_client_ = None
|
21
|
+
_async_client_ = None
|
22
|
+
|
23
|
+
_sync_client_instance = None
|
24
|
+
_async_client_instance = None
|
25
|
+
|
26
|
+
key_sequence = None
|
27
|
+
usage_sequence = None
|
28
|
+
model_exclude_list = []
|
29
|
+
input_token_name = "prompt_tokens"
|
30
|
+
output_token_name = "completion_tokens"
|
31
|
+
|
32
|
+
@classmethod
|
33
|
+
def available(cls) -> list[str]:
|
34
|
+
return ["test"]
|
35
|
+
|
36
|
+
@classmethod
|
37
|
+
def create_model(cls, model_name, model_class_name=None) -> LanguageModel:
|
38
|
+
throw_exception = False
|
39
|
+
|
40
|
+
class TestServiceLanguageModel(LanguageModel):
|
41
|
+
_model_ = "test"
|
42
|
+
_parameters_ = {"temperature": 0.5}
|
43
|
+
_inference_service_ = InferenceServiceType.TEST.value
|
44
|
+
usage_sequence = ["usage"]
|
45
|
+
key_sequence = ["message", 0, "text"]
|
46
|
+
input_token_name = cls.input_token_name
|
47
|
+
output_token_name = cls.output_token_name
|
48
|
+
_rpm = 1000
|
49
|
+
_tpm = 100000
|
50
|
+
|
51
|
+
@property
|
52
|
+
def _canned_response(self):
|
53
|
+
if hasattr(self, "canned_response"):
|
54
|
+
|
55
|
+
return self.canned_response
|
56
|
+
else:
|
57
|
+
return "Hello, world"
|
58
|
+
|
59
|
+
async def async_execute_model_call(
|
60
|
+
self,
|
61
|
+
user_prompt: str,
|
62
|
+
system_prompt: str,
|
63
|
+
# func: Optional[callable] = None,
|
64
|
+
files_list: Optional[List["File"]] = None,
|
65
|
+
) -> dict[str, Any]:
|
66
|
+
await asyncio.sleep(0.1)
|
67
|
+
|
68
|
+
if hasattr(self, "throw_exception") and self.throw_exception:
|
69
|
+
if hasattr(self, "exception_probability"):
|
70
|
+
p = self.exception_probability
|
71
|
+
else:
|
72
|
+
p = 1
|
73
|
+
|
74
|
+
if random.random() < p:
|
75
|
+
raise Exception("This is a test error")
|
76
|
+
|
77
|
+
if hasattr(self, "func"):
|
78
|
+
return {
|
79
|
+
"message": [
|
80
|
+
{"text": self.func(user_prompt, system_prompt, files_list)}
|
81
|
+
],
|
82
|
+
"usage": {"prompt_tokens": 1, "completion_tokens": 1},
|
83
|
+
}
|
84
|
+
|
85
|
+
return {
|
86
|
+
"message": [{"text": f"{self._canned_response}"}],
|
87
|
+
"usage": {"prompt_tokens": 1, "completion_tokens": 1},
|
88
|
+
}
|
89
|
+
|
90
|
+
return TestServiceLanguageModel
|