eval-ai-library 0.1.0__py3-none-any.whl → 0.2.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of eval-ai-library might be problematic. Click here for more details.

@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: eval-ai-library
3
- Version: 0.1.0
3
+ Version: 0.2.1
4
4
  Summary: Comprehensive AI Model Evaluation Framework with support for multiple LLM providers
5
5
  Author-email: Aleksandr Meshkov <alekslynx90@gmail.com>
6
6
  License: MIT
@@ -1,5 +1,5 @@
1
- eval_ai_library-0.1.0.dist-info/licenses/LICENSE,sha256=rK9uLDgWNrCHNdp-Zma_XghDE7Fs0u0kDi3WMcmYx6w,1074
2
- eval_lib/__init__.py,sha256=vwyPEMhyQqJR8_usT-LzI1xjvFTBa8S16s53frz4JJM,2546
1
+ eval_ai_library-0.2.1.dist-info/licenses/LICENSE,sha256=rK9uLDgWNrCHNdp-Zma_XghDE7Fs0u0kDi3WMcmYx6w,1074
2
+ eval_lib/__init__.py,sha256=TWK-1gpyXRLYxFMB-7f30wzr4c2uXE25YgFOMKfwuDo,2533
3
3
  eval_lib/evaluate.py,sha256=XwtNkEaVzbrB2m6rgRaeFcb9kgHeHjAMM-WxMGCNs8E,12516
4
4
  eval_lib/evaluation_schema.py,sha256=7IDd_uozqewhh7k0p1hKut_20udvRxxkV6thclxKUg0,1904
5
5
  eval_lib/llm_client.py,sha256=y-6blKiMtZcPwRGeQ-_XHMIcAtgI-JxH3sArGxYErbI,7620
@@ -28,7 +28,7 @@ eval_lib/metrics/faithfulness_metric/faithfulness.py,sha256=jHZMYflkXMEDsh8czENP
28
28
  eval_lib/metrics/geval/geval.py,sha256=MhWd4KpSIsTEvanscuLsIXOtjOGEtw_2-8ctrqabzTo,10490
29
29
  eval_lib/metrics/restricted_refusal_metric/restricted_refusal.py,sha256=TQpoiN1hG2KKw1ILpJsmoE2T8uMGhxHz3wswDaqcBPA,4156
30
30
  eval_lib/metrics/toxicity_metric/toxicity.py,sha256=Rh5UPSeAtkGopmCU4JGTAVLHbgFKrhVO83pl6vdNdTw,3882
31
- eval_ai_library-0.1.0.dist-info/METADATA,sha256=dJ2lV7VKlvVfmcIimgIAfgi900g21DBayknvs5Ng8yg,23834
32
- eval_ai_library-0.1.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
33
- eval_ai_library-0.1.0.dist-info/top_level.txt,sha256=uQHpEd2XI0oZgq1eCww9zMvVgDJgwXMWkCD45fYUzEg,9
34
- eval_ai_library-0.1.0.dist-info/RECORD,,
31
+ eval_ai_library-0.2.1.dist-info/METADATA,sha256=hIrn1V-wz4UbmnHI4-81KQ6jB-Jvt00D55GnCsLilbk,23834
32
+ eval_ai_library-0.2.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
33
+ eval_ai_library-0.2.1.dist-info/top_level.txt,sha256=uQHpEd2XI0oZgq1eCww9zMvVgDJgwXMWkCD45fYUzEg,9
34
+ eval_ai_library-0.2.1.dist-info/RECORD,,
eval_lib/__init__.py CHANGED
@@ -10,7 +10,7 @@ __author__ = "Aleksandr Meskov"
10
10
 
11
11
  # Core evaluation functions
12
12
  from eval_lib.evaluate import evaluate, evaluate_conversations
13
- from eval_lib.utils import score_agg, softmax_agg
13
+ from eval_lib.utils import score_agg
14
14
 
15
15
  # Test case schemas
16
16
  from eval_lib.testcases_schema import (