ragbits-evaluate 1.3.0.dev202509120609__py3-none-any.whl → 1.4.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ragbits/evaluate/metrics/question_answer.py +25 -8
- {ragbits_evaluate-1.3.0.dev202509120609.dist-info → ragbits_evaluate-1.4.0.dist-info}/METADATA +2 -2
- {ragbits_evaluate-1.3.0.dev202509120609.dist-info → ragbits_evaluate-1.4.0.dist-info}/RECORD +4 -4
- {ragbits_evaluate-1.3.0.dev202509120609.dist-info → ragbits_evaluate-1.4.0.dist-info}/WHEEL +1 -1
|
@@ -4,14 +4,6 @@ from asyncio import AbstractEventLoop
|
|
|
4
4
|
from itertools import chain
|
|
5
5
|
from typing import Generic, TypeVar
|
|
6
6
|
|
|
7
|
-
from continuous_eval.llm_factory import LLMInterface
|
|
8
|
-
from continuous_eval.metrics.base import LLMBasedMetric
|
|
9
|
-
from continuous_eval.metrics.generation.text import (
|
|
10
|
-
LLMBasedAnswerCorrectness,
|
|
11
|
-
LLMBasedAnswerRelevance,
|
|
12
|
-
LLMBasedFaithfulness,
|
|
13
|
-
LLMBasedStyleConsistency,
|
|
14
|
-
)
|
|
15
7
|
from typing_extensions import Self
|
|
16
8
|
|
|
17
9
|
from ragbits.agents.types import QuestionAnswerPromptOutputT
|
|
@@ -20,6 +12,31 @@ from ragbits.core.utils.helpers import batched
|
|
|
20
12
|
from ragbits.evaluate.metrics.base import Metric
|
|
21
13
|
from ragbits.evaluate.pipelines.question_answer import QuestionAnswerResult
|
|
22
14
|
|
|
15
|
+
try:
|
|
16
|
+
from continuous_eval.llm_factory import LLMInterface
|
|
17
|
+
from continuous_eval.metrics.base import LLMBasedMetric
|
|
18
|
+
from continuous_eval.metrics.generation.text import (
|
|
19
|
+
LLMBasedAnswerCorrectness,
|
|
20
|
+
LLMBasedAnswerRelevance,
|
|
21
|
+
LLMBasedFaithfulness,
|
|
22
|
+
LLMBasedStyleConsistency,
|
|
23
|
+
)
|
|
24
|
+
except ModuleNotFoundError:
|
|
25
|
+
from continuous_eval.llms.base import LLMInterface
|
|
26
|
+
from continuous_eval.metrics import Metric as LLMBasedMetric
|
|
27
|
+
from continuous_eval.metrics.generation.text import (
|
|
28
|
+
AnswerCorrectness as LLMBasedAnswerCorrectness,
|
|
29
|
+
)
|
|
30
|
+
from continuous_eval.metrics.generation.text import (
|
|
31
|
+
AnswerRelevance as LLMBasedAnswerRelevance,
|
|
32
|
+
)
|
|
33
|
+
from continuous_eval.metrics.generation.text import (
|
|
34
|
+
Faithfulness as LLMBasedFaithfulness,
|
|
35
|
+
)
|
|
36
|
+
from continuous_eval.metrics.generation.text import (
|
|
37
|
+
StyleConsistency as LLMBasedStyleConsistency,
|
|
38
|
+
)
|
|
39
|
+
|
|
23
40
|
MetricT = TypeVar("MetricT", bound=LLMBasedMetric)
|
|
24
41
|
|
|
25
42
|
|
{ragbits_evaluate-1.3.0.dev202509120609.dist-info → ragbits_evaluate-1.4.0.dist-info}/METADATA
RENAMED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: ragbits-evaluate
|
|
3
|
-
Version: 1.
|
|
3
|
+
Version: 1.4.0
|
|
4
4
|
Summary: Evaluation module for Ragbits components
|
|
5
5
|
Project-URL: Homepage, https://github.com/deepsense-ai/ragbits
|
|
6
6
|
Project-URL: Bug Reports, https://github.com/deepsense-ai/ragbits/issues
|
|
@@ -27,7 +27,7 @@ Requires-Dist: distilabel<2.0.0,>=1.5.0
|
|
|
27
27
|
Requires-Dist: hydra-core<2.0.0,>=1.3.2
|
|
28
28
|
Requires-Dist: neptune[optuna]<2.0.0,>=1.12.0
|
|
29
29
|
Requires-Dist: optuna<5.0.0,>=4.0.0
|
|
30
|
-
Requires-Dist: ragbits-core==1.
|
|
30
|
+
Requires-Dist: ragbits-core==1.4.0
|
|
31
31
|
Provides-Extra: relari
|
|
32
32
|
Requires-Dist: continuous-eval<1.0.0,>=0.3.12; extra == 'relari'
|
|
33
33
|
Description-Content-Type: text/markdown
|
{ragbits_evaluate-1.3.0.dev202509120609.dist-info → ragbits_evaluate-1.4.0.dist-info}/RECORD
RENAMED
|
@@ -28,11 +28,11 @@ ragbits/evaluate/factories/__init__.py,sha256=7nh0J80EfqMWRGtHx4hkfHNMztfC6FMhH8
|
|
|
28
28
|
ragbits/evaluate/metrics/__init__.py,sha256=Mr83ytGyvdXtBlr7Bbo0-5auE0530xsd3wffKSIf8cE,95
|
|
29
29
|
ragbits/evaluate/metrics/base.py,sha256=bOscQ_nJXLGWmP2ls9jncrUoeghNBnKDJsab71pFEjo,2519
|
|
30
30
|
ragbits/evaluate/metrics/document_search.py,sha256=MfvMwEPenqiJdKYuW6WLvmtMch9ZVYb0T6ibpOF3vGI,3189
|
|
31
|
-
ragbits/evaluate/metrics/question_answer.py,sha256=
|
|
31
|
+
ragbits/evaluate/metrics/question_answer.py,sha256=4rbJ9z_eAyk_5mOeQGyaP_rqodS8MZZltwmPmDitb_4,7722
|
|
32
32
|
ragbits/evaluate/pipelines/__init__.py,sha256=Bqp_L7aRq12Ua19ELZDsdYvra6-GlLrQ9cIG2IWArko,1294
|
|
33
33
|
ragbits/evaluate/pipelines/base.py,sha256=QV3fjPnbJjeCgcbt8yV1Ho3BamEUc3wSca3MAzaBlV0,1739
|
|
34
34
|
ragbits/evaluate/pipelines/document_search.py,sha256=tgk-I21eshdBbWVsuNa1zWK_fWuDNXhhMCn1_Fdu_Ko,3840
|
|
35
35
|
ragbits/evaluate/pipelines/question_answer.py,sha256=3CYVHDLnOy4z7kgYPMluiJ8POulHo-w3PEiqvqsF4Dc,2797
|
|
36
|
-
ragbits_evaluate-1.
|
|
37
|
-
ragbits_evaluate-1.
|
|
38
|
-
ragbits_evaluate-1.
|
|
36
|
+
ragbits_evaluate-1.4.0.dist-info/METADATA,sha256=fPzCfoI9hWL9qHnOcNi-RBTMAa6rUPdeQra8uwne_so,2298
|
|
37
|
+
ragbits_evaluate-1.4.0.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
|
|
38
|
+
ragbits_evaluate-1.4.0.dist-info/RECORD,,
|