orchestrator-core 4.6.3rc1__py3-none-any.whl → 4.6.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
orchestrator/__init__.py CHANGED
@@ -13,7 +13,7 @@
13
13
 
14
14
  """This is the orchestrator workflow engine."""
15
15
 
16
- __version__ = "4.6.3rc1"
16
+ __version__ = "4.6.4"
17
17
 
18
18
 
19
19
  from structlog import get_logger
@@ -14,6 +14,7 @@ def subscriptions_command(
14
14
  subscription_id: str | None = typer.Option(None, help="UUID (default = all)"),
15
15
  dry_run: bool = typer.Option(False, help="No DB writes"),
16
16
  force_index: bool = typer.Option(False, help="Force re-index (ignore hash cache)"),
17
+ show_progress: bool = typer.Option(False, help="Show per-entity progress"),
17
18
  ) -> None:
18
19
  """Index subscription_search_index."""
19
20
  run_indexing_for_entity(
@@ -21,6 +22,7 @@ def subscriptions_command(
21
22
  entity_id=subscription_id,
22
23
  dry_run=dry_run,
23
24
  force_index=force_index,
25
+ show_progress=show_progress,
24
26
  )
25
27
 
26
28
 
@@ -29,6 +31,7 @@ def products_command(
29
31
  product_id: str | None = typer.Option(None, help="UUID (default = all)"),
30
32
  dry_run: bool = typer.Option(False, help="No DB writes"),
31
33
  force_index: bool = typer.Option(False, help="Force re-index (ignore hash cache)"),
34
+ show_progress: bool = typer.Option(False, help="Show per-entity progress"),
32
35
  ) -> None:
33
36
  """Index product_search_index."""
34
37
  run_indexing_for_entity(
@@ -36,6 +39,7 @@ def products_command(
36
39
  entity_id=product_id,
37
40
  dry_run=dry_run,
38
41
  force_index=force_index,
42
+ show_progress=show_progress,
39
43
  )
40
44
 
41
45
 
@@ -44,6 +48,7 @@ def processes_command(
44
48
  process_id: str | None = typer.Option(None, help="UUID (default = all)"),
45
49
  dry_run: bool = typer.Option(False, help="No DB writes"),
46
50
  force_index: bool = typer.Option(False, help="Force re-index (ignore hash cache)"),
51
+ show_progress: bool = typer.Option(False, help="Show per-entity progress"),
47
52
  ) -> None:
48
53
  """Index process_search_index."""
49
54
  run_indexing_for_entity(
@@ -51,6 +56,7 @@ def processes_command(
51
56
  entity_id=process_id,
52
57
  dry_run=dry_run,
53
58
  force_index=force_index,
59
+ show_progress=show_progress,
54
60
  )
55
61
 
56
62
 
@@ -59,6 +65,7 @@ def workflows_command(
59
65
  workflow_id: str | None = typer.Option(None, help="UUID (default = all)"),
60
66
  dry_run: bool = typer.Option(False, help="No DB writes"),
61
67
  force_index: bool = typer.Option(False, help="Force re-index (ignore hash cache)"),
68
+ show_progress: bool = typer.Option(False, help="Show per-entity progress"),
62
69
  ) -> None:
63
70
  """Index workflow_search_index."""
64
71
  run_indexing_for_entity(
@@ -66,6 +73,7 @@ def workflows_command(
66
73
  entity_id=workflow_id,
67
74
  dry_run=dry_run,
68
75
  force_index=force_index,
76
+ show_progress=show_progress,
69
77
  )
70
78
 
71
79
 
@@ -1,11 +1,17 @@
1
1
  from collections.abc import Sequence
2
+ from functools import wraps
3
+ from typing import Callable, Coroutine
2
4
 
5
+ import structlog
3
6
  from sqlalchemy import CompoundSelect, Select, select
4
7
  from sqlalchemy.orm.strategy_options import _AbstractLoad
8
+ from starlette.concurrency import run_in_threadpool
5
9
 
6
10
  from orchestrator.db import db
7
11
  from orchestrator.db.database import BaseModel
8
12
 
13
+ logger = structlog.get_logger(__name__)
14
+
9
15
 
10
16
  def rows_from_statement(
11
17
  stmt: Select | CompoundSelect,
@@ -19,3 +25,12 @@ def rows_from_statement(
19
25
  result = db.session.scalars(from_stmt)
20
26
  uresult = result.unique() if unique else result
21
27
  return uresult.all()
28
+
29
+
30
+ def make_async(f: Callable): # type: ignore
31
+ @wraps(f)
32
+ async def wrapper(*args, **kwargs) -> Coroutine: # type: ignore
33
+ logger.debug(f"**async, calling fn {f.__name__}")
34
+ return await run_in_threadpool(f, *args, **kwargs)
35
+
36
+ return wrapper
@@ -25,7 +25,7 @@ from orchestrator.db.range import apply_range_to_statement
25
25
  from orchestrator.db.sorting import Sort
26
26
  from orchestrator.db.sorting.process import process_sort_fields, sort_processes
27
27
  from orchestrator.graphql.pagination import Connection
28
- from orchestrator.graphql.resolvers.helpers import rows_from_statement
28
+ from orchestrator.graphql.resolvers.helpers import make_async, rows_from_statement
29
29
  from orchestrator.graphql.schemas.process import ProcessType
30
30
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
31
31
  from orchestrator.graphql.utils import (
@@ -55,7 +55,8 @@ def _enrich_process(process: ProcessTable, with_details: bool = False) -> Proces
55
55
  return ProcessSchema(**process_data)
56
56
 
57
57
 
58
- async def resolve_process(info: OrchestratorInfo, process_id: UUID) -> ProcessType | None:
58
+ @make_async
59
+ def resolve_process(info: OrchestratorInfo, process_id: UUID) -> ProcessType | None:
59
60
  query_loaders = get_query_loaders_for_gql_fields(ProcessTable, info)
60
61
  stmt = select(ProcessTable).options(*query_loaders).where(ProcessTable.process_id == process_id)
61
62
  if process := db.session.scalar(stmt):
@@ -64,7 +65,8 @@ async def resolve_process(info: OrchestratorInfo, process_id: UUID) -> ProcessTy
64
65
  return None
65
66
 
66
67
 
67
- async def resolve_processes(
68
+ @make_async
69
+ def resolve_processes(
68
70
  info: OrchestratorInfo,
69
71
  filter_by: list[GraphqlFilter] | None = None,
70
72
  sort_by: list[GraphqlSort] | None = None,
@@ -9,7 +9,7 @@ from orchestrator.db.range.range import apply_range_to_statement
9
9
  from orchestrator.db.sorting import Sort
10
10
  from orchestrator.db.sorting.product import product_sort_fields, sort_products
11
11
  from orchestrator.graphql.pagination import Connection
12
- from orchestrator.graphql.resolvers.helpers import rows_from_statement
12
+ from orchestrator.graphql.resolvers.helpers import make_async, rows_from_statement
13
13
  from orchestrator.graphql.schemas.product import ProductType
14
14
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
15
15
  from orchestrator.graphql.utils import create_resolver_error_handler, is_querying_page_data, to_graphql_result_page
@@ -19,7 +19,8 @@ from orchestrator.utils.search_query import create_sqlalchemy_select
19
19
  logger = structlog.get_logger(__name__)
20
20
 
21
21
 
22
- async def resolve_products(
22
+ @make_async
23
+ def resolve_products(
23
24
  info: OrchestratorInfo,
24
25
  filter_by: list[GraphqlFilter] | None = None,
25
26
  sort_by: list[GraphqlSort] | None = None,
@@ -13,7 +13,7 @@ from orchestrator.db.range.range import apply_range_to_statement
13
13
  from orchestrator.db.sorting import Sort
14
14
  from orchestrator.db.sorting.product_block import product_block_sort_fields, sort_product_blocks
15
15
  from orchestrator.graphql.pagination import Connection
16
- from orchestrator.graphql.resolvers.helpers import rows_from_statement
16
+ from orchestrator.graphql.resolvers.helpers import make_async, rows_from_statement
17
17
  from orchestrator.graphql.schemas.product_block import ProductBlock
18
18
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
19
19
  from orchestrator.graphql.utils import create_resolver_error_handler, is_querying_page_data, to_graphql_result_page
@@ -23,7 +23,8 @@ from orchestrator.utils.search_query import create_sqlalchemy_select
23
23
  logger = structlog.get_logger(__name__)
24
24
 
25
25
 
26
- async def resolve_product_blocks(
26
+ @make_async
27
+ def resolve_product_blocks(
27
28
  info: OrchestratorInfo,
28
29
  filter_by: list[GraphqlFilter] | None = None,
29
30
  sort_by: list[GraphqlSort] | None = None,
@@ -13,7 +13,7 @@ from orchestrator.db.range import apply_range_to_statement
13
13
  from orchestrator.db.sorting import Sort
14
14
  from orchestrator.db.sorting.resource_type import resource_type_sort_fields, sort_resource_types
15
15
  from orchestrator.graphql.pagination import Connection
16
- from orchestrator.graphql.resolvers.helpers import rows_from_statement
16
+ from orchestrator.graphql.resolvers.helpers import make_async, rows_from_statement
17
17
  from orchestrator.graphql.schemas.resource_type import ResourceType
18
18
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
19
19
  from orchestrator.graphql.utils import create_resolver_error_handler, is_querying_page_data, to_graphql_result_page
@@ -23,7 +23,8 @@ from orchestrator.utils.search_query import create_sqlalchemy_select
23
23
  logger = structlog.get_logger(__name__)
24
24
 
25
25
 
26
- async def resolve_resource_types(
26
+ @make_async
27
+ def resolve_resource_types(
27
28
  info: OrchestratorInfo,
28
29
  filter_by: list[GraphqlFilter] | None = None,
29
30
  sort_by: list[GraphqlSort] | None = None,
@@ -3,6 +3,7 @@ import structlog
3
3
  from orchestrator.db.filters import Filter
4
4
  from orchestrator.db.sorting import Sort
5
5
  from orchestrator.graphql.pagination import Connection
6
+ from orchestrator.graphql.resolvers.helpers import make_async
6
7
  from orchestrator.graphql.schemas.scheduled_task import ScheduledTaskGraphql
7
8
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
8
9
  from orchestrator.graphql.utils import create_resolver_error_handler, to_graphql_result_page
@@ -12,7 +13,8 @@ from orchestrator.schedules.scheduler import get_scheduler_tasks, scheduled_task
12
13
  logger = structlog.get_logger(__name__)
13
14
 
14
15
 
15
- async def resolve_scheduled_tasks(
16
+ @make_async
17
+ def resolve_scheduled_tasks(
16
18
  info: OrchestratorInfo,
17
19
  filter_by: list[GraphqlFilter] | None = None,
18
20
  sort_by: list[GraphqlSort] | None = None,
@@ -4,6 +4,7 @@ from redis.asyncio import Redis as AIORedis
4
4
 
5
5
  from oauth2_lib.strawberry import authenticated_mutation_field
6
6
  from orchestrator.api.api_v1.endpoints.settings import generate_engine_status_response
7
+ from orchestrator.graphql.resolvers.helpers import make_async
7
8
  from orchestrator.graphql.schemas.errors import Error
8
9
  from orchestrator.graphql.schemas.settings import (
9
10
  CACHE_FLUSH_OPTIONS,
@@ -27,6 +28,7 @@ logger = structlog.get_logger(__name__)
27
28
 
28
29
 
29
30
  # Queries
31
+ @make_async
30
32
  def resolve_settings(info: OrchestratorInfo) -> StatusType:
31
33
  selected_fields = get_selected_fields(info)
32
34
 
@@ -18,6 +18,7 @@ from graphql import GraphQLError
18
18
  from pydantic.alias_generators import to_camel as to_lower_camel
19
19
  from sqlalchemy import Select, func, select
20
20
  from sqlalchemy.orm import contains_eager
21
+ from starlette.concurrency import run_in_threadpool
21
22
  from strawberry.experimental.pydantic.conversion_types import StrawberryTypeFromPydantic
22
23
 
23
24
  from nwastdlib.asyncio import gather_nice
@@ -101,7 +102,7 @@ async def format_subscription(info: OrchestratorInfo, subscription: Subscription
101
102
  async def resolve_subscription(info: OrchestratorInfo, id: UUID) -> SubscriptionInterface | None:
102
103
  stmt = select(SubscriptionTable).where(SubscriptionTable.subscription_id == id)
103
104
 
104
- if subscription := db.session.scalar(stmt):
105
+ if subscription := await run_in_threadpool(db.session.scalar, stmt):
105
106
  return await format_subscription(info, subscription)
106
107
  return None
107
108
 
@@ -141,12 +142,13 @@ async def resolve_subscriptions(
141
142
  stmt = filter_by_query_string(stmt, query)
142
143
 
143
144
  stmt = cast(Select, sort_subscriptions(stmt, pydantic_sort_by, _error_handler))
144
- total = db.session.scalar(select(func.count()).select_from(stmt.subquery()))
145
+ total = await run_in_threadpool(db.session.scalar, select(func.count()).select_from(stmt.subquery()))
145
146
  stmt = apply_range_to_statement(stmt, after, after + first + 1)
146
147
 
147
148
  graphql_subscriptions: list[SubscriptionInterface] = []
148
149
  if is_querying_page_data(info):
149
- subscriptions = db.session.scalars(stmt).all()
150
+ scalars = await run_in_threadpool(db.session.scalars, stmt)
151
+ subscriptions = scalars.all()
150
152
  graphql_subscriptions = list(await gather_nice((format_subscription(info, p) for p in subscriptions))) # type: ignore
151
153
  logger.info("Resolve subscriptions", filter_by=filter_by, total=total)
152
154
 
@@ -1,6 +1,7 @@
1
1
  from structlog import get_logger
2
2
 
3
3
  from orchestrator import __version__
4
+ from orchestrator.graphql.resolvers.helpers import make_async
4
5
  from orchestrator.graphql.schemas.version import VersionType
5
6
  from orchestrator.graphql.types import OrchestratorInfo
6
7
  from orchestrator.graphql.utils import create_resolver_error_handler
@@ -11,6 +12,7 @@ logger = get_logger(__name__)
11
12
  VERSIONS = [f"orchestrator-core: {__version__}"]
12
13
 
13
14
 
15
+ @make_async
14
16
  def resolve_version(info: OrchestratorInfo) -> VersionType | None:
15
17
  logger.debug("resolve_version() called")
16
18
  _error_handler = create_resolver_error_handler(info)
@@ -9,7 +9,7 @@ from orchestrator.db.range.range import apply_range_to_statement
9
9
  from orchestrator.db.sorting import Sort
10
10
  from orchestrator.db.sorting.workflow import sort_workflows, workflow_sort_fields
11
11
  from orchestrator.graphql.pagination import Connection
12
- from orchestrator.graphql.resolvers.helpers import rows_from_statement
12
+ from orchestrator.graphql.resolvers.helpers import make_async, rows_from_statement
13
13
  from orchestrator.graphql.schemas.workflow import Workflow
14
14
  from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
15
15
  from orchestrator.graphql.utils import create_resolver_error_handler, is_querying_page_data, to_graphql_result_page
@@ -19,7 +19,8 @@ from orchestrator.utils.search_query import create_sqlalchemy_select
19
19
  logger = structlog.get_logger(__name__)
20
20
 
21
21
 
22
- async def resolve_workflows(
22
+ @make_async
23
+ def resolve_workflows(
23
24
  info: OrchestratorInfo,
24
25
  filter_by: list[GraphqlFilter] | None = None,
25
26
  sort_by: list[GraphqlSort] | None = None,
@@ -14,7 +14,6 @@
14
14
  from datetime import date, datetime
15
15
  from typing import Annotated, Any, Literal
16
16
 
17
- from dateutil.parser import parse as dt_parse
18
17
  from pydantic import BaseModel, BeforeValidator, Field, model_validator
19
18
  from sqlalchemy import TIMESTAMP, and_
20
19
  from sqlalchemy import cast as sa_cast
@@ -27,10 +26,10 @@ def _validate_date_string(v: Any) -> Any:
27
26
  if not isinstance(v, str):
28
27
  return v
29
28
  try:
30
- dt_parse(v)
29
+ datetime.fromisoformat(v)
31
30
  return v
32
31
  except Exception as exc:
33
- raise ValueError("is not a valid date or datetime string") from exc
32
+ raise ValueError("is not a valid ISO-8601 date or datetime string") from exc
34
33
 
35
34
 
36
35
  DateValue = datetime | date | str
@@ -44,8 +43,8 @@ class DateRange(BaseModel):
44
43
 
45
44
  @model_validator(mode="after")
46
45
  def _order(self) -> "DateRange":
47
- to_datetime = dt_parse(str(self.end))
48
- from_datetime = dt_parse(str(self.start))
46
+ to_datetime = datetime.fromisoformat(str(self.end))
47
+ from_datetime = datetime.fromisoformat(str(self.start))
49
48
  if to_datetime <= from_datetime:
50
49
  raise ValueError("'to' must be after 'from'")
51
50
  return self
@@ -45,6 +45,23 @@ def _maybe_begin(session: Session | None) -> Iterator[None]:
45
45
  yield
46
46
 
47
47
 
48
+ @contextmanager
49
+ def _maybe_progress(show_progress: bool, total_count: int | None, label: str) -> Iterator[Any]:
50
+ """Context manager that optionally creates a progress bar."""
51
+ if show_progress:
52
+ import typer
53
+
54
+ with typer.progressbar(
55
+ length=total_count,
56
+ label=label,
57
+ show_eta=True,
58
+ show_percent=bool(total_count),
59
+ ) as progress:
60
+ yield progress
61
+ else:
62
+ yield None
63
+
64
+
48
65
  class Indexer:
49
66
  """Index entities into `AiSearchIndex` using streaming reads and batched writes.
50
67
 
@@ -89,11 +106,21 @@ class Indexer:
89
106
  8) Repeat until the stream is exhausted.
90
107
  """
91
108
 
92
- def __init__(self, config: EntityConfig, dry_run: bool, force_index: bool, chunk_size: int = 1000) -> None:
109
+ def __init__(
110
+ self,
111
+ config: EntityConfig,
112
+ dry_run: bool,
113
+ force_index: bool,
114
+ chunk_size: int = 1000,
115
+ show_progress: bool = False,
116
+ total_count: int | None = None,
117
+ ) -> None:
93
118
  self.config = config
94
119
  self.dry_run = dry_run
95
120
  self.force_index = force_index
96
121
  self.chunk_size = chunk_size
122
+ self.show_progress = show_progress
123
+ self.total_count = total_count
97
124
  self.embedding_model = llm_settings.EMBEDDING_MODEL
98
125
  self.logger = logger.bind(entity_kind=config.entity_kind.value)
99
126
  self._entity_titles: dict[str, str] = {}
@@ -116,13 +143,22 @@ class Indexer:
116
143
 
117
144
  with write_scope as database:
118
145
  session: Session | None = getattr(database, "session", None)
119
- for entity in entities:
120
- chunk.append(entity)
121
- if len(chunk) >= self.chunk_size:
122
- flush()
123
146
 
124
- if chunk:
125
- flush()
147
+ with _maybe_progress(
148
+ self.show_progress, self.total_count, f"Indexing {self.config.entity_kind.value}"
149
+ ) as progress:
150
+ for entity in entities:
151
+ chunk.append(entity)
152
+
153
+ if len(chunk) >= self.chunk_size:
154
+ flush()
155
+ if progress:
156
+ progress.update(self.chunk_size)
157
+
158
+ if chunk:
159
+ flush()
160
+ if progress:
161
+ progress.update(len(chunk))
126
162
 
127
163
  final_log_message = (
128
164
  f"processed {total_records_processed} records and skipped {total_identical_records} identical records."
@@ -66,6 +66,21 @@ class EntityConfig(Generic[ModelT]):
66
66
  return "UNKNOWN"
67
67
 
68
68
 
69
+ @dataclass(frozen=True)
70
+ class ProcessConfig(EntityConfig[ProcessTable]):
71
+ """Processes need to eager load workflow for workflow_name field."""
72
+
73
+ def get_all_query(self, entity_id: str | None = None) -> Query | Select:
74
+ from sqlalchemy.orm import selectinload
75
+
76
+ # Only load workflow, not subscriptions (keeps it lightweight)
77
+ query = self.table.query.options(selectinload(ProcessTable.workflow))
78
+ if entity_id:
79
+ pk_column = getattr(self.table, self.pk_name)
80
+ query = query.filter(pk_column == UUID(entity_id))
81
+ return query
82
+
83
+
69
84
  @dataclass(frozen=True)
70
85
  class WorkflowConfig(EntityConfig[WorkflowTable]):
71
86
  """Workflows have a custom select() function that filters out deleted workflows."""
@@ -95,7 +110,7 @@ ENTITY_CONFIG_REGISTRY: dict[EntityType, EntityConfig] = {
95
110
  root_name="product",
96
111
  title_paths=["product.description", "product.name"],
97
112
  ),
98
- EntityType.PROCESS: EntityConfig(
113
+ EntityType.PROCESS: ProcessConfig(
99
114
  entity_kind=EntityType.PROCESS,
100
115
  table=ProcessTable,
101
116
  traverser=ProcessTraverser,
@@ -11,7 +11,10 @@
11
11
  # See the License for the specific language governing permissions and
12
12
  # limitations under the License.
13
13
 
14
+ from typing import Any
15
+
14
16
  import structlog
17
+ from sqlalchemy import func, select
15
18
  from sqlalchemy.orm import Query
16
19
 
17
20
  from orchestrator.db import db
@@ -23,12 +26,20 @@ from orchestrator.search.indexing.registry import ENTITY_CONFIG_REGISTRY
23
26
  logger = structlog.get_logger(__name__)
24
27
 
25
28
 
29
+ def _get_entity_count(stmt: Any) -> int | None:
30
+ """Get total count of entities from a select statement."""
31
+
32
+ count_stmt = select(func.count()).select_from(stmt.subquery())
33
+ return db.session.execute(count_stmt).scalar()
34
+
35
+
26
36
  def run_indexing_for_entity(
27
37
  entity_kind: EntityType,
28
38
  entity_id: str | None = None,
29
39
  dry_run: bool = False,
30
40
  force_index: bool = False,
31
41
  chunk_size: int = 1000,
42
+ show_progress: bool = False,
32
43
  ) -> None:
33
44
  """Stream and index entities for the given kind.
34
45
 
@@ -46,6 +57,7 @@ def run_indexing_for_entity(
46
57
  existing hashes.
47
58
  chunk_size (int): Number of rows fetched per round-trip and passed to
48
59
  the indexer per batch.
60
+ show_progress (bool): When True, logs progress for each processed entity.
49
61
 
50
62
  Returns:
51
63
  None
@@ -60,10 +72,19 @@ def run_indexing_for_entity(
60
72
  else:
61
73
  stmt = q
62
74
 
75
+ total_count = _get_entity_count(stmt) if show_progress else None
76
+
63
77
  stmt = stmt.execution_options(stream_results=True, yield_per=chunk_size)
64
78
  entities = db.session.execute(stmt).scalars()
65
79
 
66
- indexer = Indexer(config=config, dry_run=dry_run, force_index=force_index, chunk_size=chunk_size)
80
+ indexer = Indexer(
81
+ config=config,
82
+ dry_run=dry_run,
83
+ force_index=force_index,
84
+ chunk_size=chunk_size,
85
+ show_progress=show_progress,
86
+ total_count=total_count,
87
+ )
67
88
 
68
89
  with cache_subscription_models():
69
90
  indexer.run(entities)
@@ -29,7 +29,7 @@ from orchestrator.domain.base import ProductBlockModel, ProductModel
29
29
  from orchestrator.domain.lifecycle import (
30
30
  lookup_specialized_type,
31
31
  )
32
- from orchestrator.schemas.process import ProcessSchema
32
+ from orchestrator.schemas.process import ProcessBaseSchema
33
33
  from orchestrator.schemas.workflow import WorkflowSchema
34
34
  from orchestrator.search.core.exceptions import ModelLoadError, ProductNotInRegistryError
35
35
  from orchestrator.search.core.types import LTREE_SEPARATOR, ExtractedField, FieldType
@@ -307,17 +307,39 @@ class ProductTraverser(BaseTraverser):
307
307
 
308
308
 
309
309
  class ProcessTraverser(BaseTraverser):
310
- """Traverser for process entities using ProcessSchema model.
310
+ """Traverser for process entities using ProcessBaseSchema.
311
311
 
312
- Note: Currently extracts only top-level process fields. Could be extended to include:
313
- - Related subscriptions (entity.subscriptions)
314
- - Related workflow information beyond workflow_name
312
+ Only indexes top-level process fields (no subscriptions or steps)
313
+ to keep the index size manageable.
315
314
  """
316
315
 
316
+ EXCLUDED_FIELDS = {"traceback", "failed_reason"}
317
+
318
+ @classmethod
319
+ def _load_model(cls, entity: ProcessTable) -> ProcessBaseSchema | None:
320
+ return cls._load_model_with_schema(entity, ProcessBaseSchema, "process_id")
321
+
317
322
  @classmethod
318
- def _load_model(cls, process: ProcessTable) -> ProcessSchema:
319
- """Load process model using ProcessSchema."""
320
- return cls._load_model_with_schema(process, ProcessSchema, "process_id")
323
+ def get_fields(cls, entity: ProcessTable, pk_name: str, root_name: str) -> list[ExtractedField]: # type: ignore[override]
324
+ """Extract fields from process, excluding fields in EXCLUDED_FIELDS."""
325
+ try:
326
+ model = cls._load_model(entity)
327
+ if model is None:
328
+ return []
329
+
330
+ return sorted(
331
+ (
332
+ field
333
+ for field in cls.traverse(model, root_name)
334
+ if field.path.split(LTREE_SEPARATOR)[-1] not in cls.EXCLUDED_FIELDS
335
+ ),
336
+ key=lambda f: f.path,
337
+ )
338
+
339
+ except (ProductNotInRegistryError, ModelLoadError) as e:
340
+ entity_id = getattr(entity, pk_name, "unknown")
341
+ logger.error(f"Failed to extract fields from {entity.__class__.__name__}", id=str(entity_id), error=str(e))
342
+ return []
321
343
 
322
344
 
323
345
  class WorkflowTraverser(BaseTraverser):
@@ -139,6 +139,63 @@ def format_aggregation_response(
139
139
  )
140
140
 
141
141
 
142
+ def truncate_text_with_highlights(
143
+ text: str, highlight_indices: list[tuple[int, int]] | None = None, max_length: int = 500, context_chars: int = 100
144
+ ) -> tuple[str, list[tuple[int, int]] | None]:
145
+ """Truncate text to max_length while preserving context around the first highlight.
146
+
147
+ Args:
148
+ text: The text to truncate
149
+ highlight_indices: List of (start, end) tuples indicating highlight positions, or None
150
+ max_length: Maximum length of the returned text
151
+ context_chars: Number of characters to show before and after the first highlight
152
+
153
+ Returns:
154
+ Tuple of (truncated_text, adjusted_highlight_indices)
155
+ """
156
+ # If text is short enough, return as-is
157
+ if len(text) <= max_length:
158
+ return text, highlight_indices
159
+
160
+ # If no highlights, truncate from beginning
161
+ if not highlight_indices:
162
+ truncated_text = text[:max_length]
163
+ suffix = "..." if len(text) > max_length else ""
164
+ return truncated_text + suffix, None
165
+
166
+ # Use first highlight to determine what to show
167
+ first_highlight_start = highlight_indices[0][0]
168
+
169
+ # Calculate start position: try to center around first highlight
170
+ start = max(0, first_highlight_start - context_chars)
171
+ end = min(len(text), start + max_length)
172
+
173
+ # Adjust start if we hit the end boundary
174
+ if end == len(text) and (end - start) < max_length:
175
+ start = max(0, end - max_length)
176
+
177
+ truncated_text = text[start:end]
178
+
179
+ # Add ellipsis to indicate truncation
180
+ truncated_from_start = start > 0
181
+ truncated_from_end = end < len(text)
182
+
183
+ if truncated_from_start:
184
+ truncated_text = "..." + truncated_text
185
+ if truncated_from_end:
186
+ truncated_text = truncated_text + "..."
187
+
188
+ # Adjust highlight indices to be relative to truncated text
189
+ offset = start - (3 if truncated_from_start else 0) # Account for leading "..."
190
+ adjusted_indices = []
191
+ for hl_start, hl_end in highlight_indices:
192
+ # Only include highlights that are within the truncated range
193
+ if hl_start >= start and hl_end <= end:
194
+ adjusted_indices.append((hl_start - offset, hl_end - offset))
195
+
196
+ return truncated_text, adjusted_indices if adjusted_indices else None
197
+
198
+
142
199
  def generate_highlight_indices(text: str, term: str) -> list[tuple[int, int]]:
143
200
  """Finds all occurrences of individual words from the term, including both word boundary and substring matches."""
144
201
  import re
@@ -201,8 +258,9 @@ def format_search_response(
201
258
  if not isinstance(path, str):
202
259
  path = str(path)
203
260
 
204
- highlight_indices = generate_highlight_indices(text, user_query) or None
205
- matching_field = MatchingField(text=text, path=path, highlight_indices=highlight_indices)
261
+ highlight_indices = generate_highlight_indices(text, user_query)
262
+ truncated_text, adjusted_indices = truncate_text_with_highlights(text, highlight_indices)
263
+ matching_field = MatchingField(text=truncated_text, path=path, highlight_indices=adjusted_indices)
206
264
 
207
265
  elif not user_query and query.filters and metadata.search_type == "structured":
208
266
  # Structured search (filter-only)
@@ -14,6 +14,7 @@
14
14
  from .base import Retriever
15
15
  from .fuzzy import FuzzyRetriever
16
16
  from .hybrid import RrfHybridRetriever
17
+ from .process import ProcessHybridRetriever
17
18
  from .semantic import SemanticRetriever
18
19
  from .structured import StructuredRetriever
19
20
 
@@ -21,6 +22,7 @@ __all__ = [
21
22
  "Retriever",
22
23
  "FuzzyRetriever",
23
24
  "RrfHybridRetriever",
25
+ "ProcessHybridRetriever",
24
26
  "SemanticRetriever",
25
27
  "StructuredRetriever",
26
28
  ]
@@ -17,7 +17,7 @@ from decimal import Decimal
17
17
  import structlog
18
18
  from sqlalchemy import BindParameter, Numeric, Select, literal
19
19
 
20
- from orchestrator.search.core.types import FieldType, SearchMetadata
20
+ from orchestrator.search.core.types import EntityType, FieldType, SearchMetadata
21
21
  from orchestrator.search.query.queries import ExportQuery, SelectQuery
22
22
 
23
23
  from ..pagination import PageCursor
@@ -63,12 +63,15 @@ class Retriever(ABC):
63
63
  Returns:
64
64
  A concrete retriever instance based on available search criteria
65
65
  """
66
+
66
67
  from .fuzzy import FuzzyRetriever
67
68
  from .hybrid import RrfHybridRetriever
69
+ from .process import ProcessHybridRetriever
68
70
  from .semantic import SemanticRetriever
69
71
  from .structured import StructuredRetriever
70
72
 
71
73
  fuzzy_term = query.fuzzy_term
74
+ is_process = query.entity_type == EntityType.PROCESS
72
75
 
73
76
  # If vector_query exists but embedding generation failed, fall back to fuzzy search with full query text
74
77
  if query_embedding is None and query.vector_query is not None and query.query_text is not None:
@@ -76,10 +79,14 @@ class Retriever(ABC):
76
79
 
77
80
  # Select retriever based on available search criteria
78
81
  if query_embedding is not None and fuzzy_term is not None:
82
+ if is_process:
83
+ return ProcessHybridRetriever(query_embedding, fuzzy_term, cursor)
79
84
  return RrfHybridRetriever(query_embedding, fuzzy_term, cursor)
80
85
  if query_embedding is not None:
81
86
  return SemanticRetriever(query_embedding, cursor)
82
87
  if fuzzy_term is not None:
88
+ if is_process:
89
+ return ProcessHybridRetriever(None, fuzzy_term, cursor)
83
90
  return FuzzyRetriever(fuzzy_term, cursor)
84
91
 
85
92
  return StructuredRetriever(cursor)
@@ -0,0 +1,225 @@
1
+ # Copyright 2019-2025 SURF, GÉANT.
2
+ # Licensed under the Apache License, Version 2.0 (the "License");
3
+ # you may not use this file except in compliance with the License.
4
+ # You may obtain a copy of the License at
5
+ #
6
+ # http://www.apache.org/licenses/LICENSE-2.0
7
+ #
8
+ # Unless required by applicable law or agreed to in writing, software
9
+ # distributed under the License is distributed on an "AS IS" BASIS,
10
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
11
+ # See the License for the specific language governing permissions and
12
+ # limitations under the License.
13
+
14
+ from typing import Any
15
+
16
+ from sqlalchemy import BindParameter, Select, String, and_, case, cast, func, literal, select
17
+ from sqlalchemy.sql.expression import ColumnElement, Label
18
+ from sqlalchemy_utils import LtreeType
19
+
20
+ from orchestrator.db.models import AiSearchIndex, ProcessStepTable
21
+ from orchestrator.search.core.types import SearchMetadata
22
+
23
+ from .hybrid import RrfHybridRetriever, compute_rrf_hybrid_score_sql
24
+
25
+
26
+ class ProcessHybridRetriever(RrfHybridRetriever):
27
+ """Process-specific hybrid retriever with process.last_step JSONB search.
28
+
29
+ Extends RrfHybridRetriever to include fuzzy search over the process.last_step
30
+ (JSONB) column. For process searches:
31
+ - Indexed fields (from AiSearchIndex): semantic + fuzzy search
32
+ - Last step JSONB field: fuzzy search only (no embeddings for dynamic data)
33
+
34
+ The retriever:
35
+ 1. Gets field candidates from AiSearchIndex
36
+ 2. Uses process.last_step JSONB column directly for fuzzy matching
37
+ 3. Combines both sources (indexed + JSONB) in unified ranking
38
+ """
39
+
40
+ q_vec: list[float] | None # type: ignore[assignment] # Override parent's type to allow None for fuzzy-only search
41
+
42
+ def __init__(self, q_vec: list[float] | None, *args: Any, **kwargs: Any) -> None:
43
+ # ProcessHybridRetriever accepts None for q_vec (fuzzy-only search)
44
+ # We pass empty list to parent to satisfy type requirements, but override behavior in _get_semantic_distance_expr
45
+ super().__init__(q_vec or [], *args, **kwargs)
46
+ self.q_vec = q_vec
47
+
48
+ def _get_semantic_distance_expr(self) -> Label[Any]:
49
+ """Get semantic distance expression, handling optional q_vec."""
50
+ if self.q_vec is None:
51
+ return literal(1.0).label("semantic_distance")
52
+
53
+ from sqlalchemy import bindparam
54
+
55
+ q_param: BindParameter[list[float]] = bindparam("q_vec", type_=AiSearchIndex.embedding.type)
56
+ sem_expr = case(
57
+ (AiSearchIndex.embedding.is_(None), None),
58
+ else_=AiSearchIndex.embedding.op("<->")(q_param),
59
+ )
60
+ return func.coalesce(sem_expr, literal(1.0)).label("semantic_distance")
61
+
62
+ def _build_indexed_candidates(
63
+ self, cand: Any, sem_val: Label[Any], best_similarity: ColumnElement[Any], filter_condition: ColumnElement[Any]
64
+ ) -> Select:
65
+ """Build candidates from indexed fields in AiSearchIndex."""
66
+ return (
67
+ select(
68
+ AiSearchIndex.entity_id,
69
+ AiSearchIndex.entity_title,
70
+ AiSearchIndex.path,
71
+ AiSearchIndex.value,
72
+ sem_val,
73
+ best_similarity.label("fuzzy_score"),
74
+ )
75
+ .select_from(AiSearchIndex)
76
+ .join(cand, cand.c.entity_id == AiSearchIndex.entity_id)
77
+ .where(
78
+ and_(
79
+ AiSearchIndex.value_type.in_(self.SEARCHABLE_FIELD_TYPES),
80
+ filter_condition,
81
+ )
82
+ )
83
+ .limit(self.field_candidates_limit)
84
+ )
85
+
86
+ def _build_jsonb_candidates(self, cand: Any) -> Select:
87
+ """Build candidates from last process_step.state JSONB column."""
88
+ # Get the last step per process using LATERAL subquery
89
+ last_step_subq = (
90
+ select(ProcessStepTable.process_id, ProcessStepTable.state)
91
+ .where(ProcessStepTable.process_id == cand.c.entity_id)
92
+ .order_by(ProcessStepTable.completed_at.desc())
93
+ .limit(1)
94
+ .lateral("last_step")
95
+ )
96
+
97
+ # Cast JSONB to text for substring search
98
+ state_text = cast(last_step_subq.c.state, String)
99
+ jsonb_fuzzy_score = func.word_similarity(self.fuzzy_term, state_text)
100
+ jsonb_filter = state_text.ilike(f"%{self.fuzzy_term}%")
101
+
102
+ return (
103
+ select(
104
+ cand.c.entity_id,
105
+ cand.c.entity_title,
106
+ cast(literal("process.last_step.state"), LtreeType).label("path"),
107
+ state_text.label("value"),
108
+ literal(1.0).label("semantic_distance"),
109
+ jsonb_fuzzy_score.label("fuzzy_score"),
110
+ )
111
+ .select_from(cand)
112
+ .join(last_step_subq, literal(True))
113
+ .where(and_(last_step_subq.c.state.isnot(None), jsonb_filter))
114
+ .limit(self.field_candidates_limit)
115
+ )
116
+
117
+ def apply(self, candidate_query: Select) -> Select:
118
+ """Apply process-specific hybrid search with process.last_step JSONB.
119
+
120
+ Args:
121
+ candidate_query: Base query returning process entity_id candidates
122
+
123
+ Returns:
124
+ Select statement with RRF scoring including last step JSONB fields
125
+ """
126
+ cand = candidate_query.subquery()
127
+
128
+ best_similarity = func.word_similarity(self.fuzzy_term, AiSearchIndex.value)
129
+ sem_val = self._get_semantic_distance_expr()
130
+ filter_condition = literal(self.fuzzy_term).op("<%")(AiSearchIndex.value)
131
+
132
+ indexed_candidates = self._build_indexed_candidates(cand, sem_val, best_similarity, filter_condition)
133
+ jsonb_candidates = self._build_jsonb_candidates(cand)
134
+
135
+ field_candidates = indexed_candidates.union_all(jsonb_candidates).cte("field_candidates")
136
+
137
+ entity_scores = (
138
+ select(
139
+ field_candidates.c.entity_id,
140
+ field_candidates.c.entity_title,
141
+ func.avg(field_candidates.c.semantic_distance).label("avg_semantic_distance"),
142
+ func.avg(field_candidates.c.fuzzy_score).label("avg_fuzzy_score"),
143
+ ).group_by(field_candidates.c.entity_id, field_candidates.c.entity_title)
144
+ ).cte("entity_scores")
145
+
146
+ entity_highlights = (
147
+ select(
148
+ field_candidates.c.entity_id,
149
+ func.first_value(field_candidates.c.value)
150
+ .over(
151
+ partition_by=field_candidates.c.entity_id,
152
+ order_by=[field_candidates.c.fuzzy_score.desc(), field_candidates.c.path.asc()],
153
+ )
154
+ .label(self.HIGHLIGHT_TEXT_LABEL),
155
+ func.first_value(field_candidates.c.path)
156
+ .over(
157
+ partition_by=field_candidates.c.entity_id,
158
+ order_by=[field_candidates.c.fuzzy_score.desc(), field_candidates.c.path.asc()],
159
+ )
160
+ .label(self.HIGHLIGHT_PATH_LABEL),
161
+ ).distinct(field_candidates.c.entity_id)
162
+ ).cte("entity_highlights")
163
+
164
+ ranked = (
165
+ select(
166
+ entity_scores.c.entity_id,
167
+ entity_scores.c.entity_title,
168
+ entity_scores.c.avg_semantic_distance,
169
+ entity_scores.c.avg_fuzzy_score,
170
+ entity_highlights.c.highlight_text,
171
+ entity_highlights.c.highlight_path,
172
+ func.dense_rank()
173
+ .over(
174
+ order_by=[entity_scores.c.avg_semantic_distance.asc().nulls_last(), entity_scores.c.entity_id.asc()]
175
+ )
176
+ .label("sem_rank"),
177
+ func.dense_rank()
178
+ .over(order_by=[entity_scores.c.avg_fuzzy_score.desc().nulls_last(), entity_scores.c.entity_id.asc()])
179
+ .label("fuzzy_rank"),
180
+ ).select_from(
181
+ entity_scores.join(entity_highlights, entity_scores.c.entity_id == entity_highlights.c.entity_id)
182
+ )
183
+ ).cte("ranked_results")
184
+
185
+ score_components = compute_rrf_hybrid_score_sql(
186
+ sem_rank_col=ranked.c.sem_rank,
187
+ fuzzy_rank_col=ranked.c.fuzzy_rank,
188
+ avg_fuzzy_score_col=ranked.c.avg_fuzzy_score,
189
+ k=self.k,
190
+ perfect_threshold=0.9,
191
+ score_numeric_type=self.SCORE_NUMERIC_TYPE,
192
+ )
193
+
194
+ perfect = score_components["perfect"]
195
+ normalized_score = score_components["normalized_score"]
196
+
197
+ score = cast(
198
+ func.round(cast(normalized_score, self.SCORE_NUMERIC_TYPE), self.SCORE_PRECISION),
199
+ self.SCORE_NUMERIC_TYPE,
200
+ ).label(self.SCORE_LABEL)
201
+
202
+ stmt = select(
203
+ ranked.c.entity_id,
204
+ ranked.c.entity_title,
205
+ score,
206
+ ranked.c.highlight_text,
207
+ ranked.c.highlight_path,
208
+ perfect.label("perfect_match"),
209
+ ).select_from(ranked)
210
+
211
+ stmt = self._apply_fused_pagination(stmt, score, ranked.c.entity_id)
212
+
213
+ stmt = stmt.order_by(
214
+ score.desc().nulls_last(),
215
+ ranked.c.entity_id.asc(),
216
+ )
217
+
218
+ if self.q_vec is not None:
219
+ stmt = stmt.params(q_vec=self.q_vec)
220
+
221
+ return stmt
222
+
223
+ @property
224
+ def metadata(self) -> SearchMetadata:
225
+ return SearchMetadata.hybrid()
@@ -6,6 +6,7 @@
6
6
  "subscription_id": "Subscription",
7
7
  "version": "Version",
8
8
  "subscription_id_info": "The subscription for this action",
9
+ "vlan_info": "The vlan id for this service port",
9
10
  "product_type": "Product Type"
10
11
  }
11
12
  },
@@ -1,9 +1,9 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: orchestrator-core
3
- Version: 4.6.3rc1
3
+ Version: 4.6.4
4
4
  Summary: This is the orchestrator workflow engine.
5
5
  Author-email: SURF <automation-beheer@surf.nl>
6
- Requires-Python: >=3.11,<3.14
6
+ Requires-Python: >=3.11,<3.15
7
7
  Description-Content-Type: text/markdown
8
8
  License-Expression: Apache-2.0
9
9
  Classifier: Development Status :: 5 - Production/Stable
@@ -20,6 +20,7 @@ Classifier: Programming Language :: Python :: 3
20
20
  Classifier: Programming Language :: Python :: 3.11
21
21
  Classifier: Programming Language :: Python :: 3.12
22
22
  Classifier: Programming Language :: Python :: 3.13
23
+ Classifier: Programming Language :: Python :: 3.14
23
24
  Classifier: Programming Language :: Python
24
25
  Classifier: Topic :: Internet :: WWW/HTTP :: HTTP Servers
25
26
  Classifier: Topic :: Internet :: WWW/HTTP
@@ -30,34 +31,34 @@ Classifier: Topic :: Software Development :: Libraries
30
31
  Classifier: Topic :: Software Development
31
32
  Classifier: Typing :: Typed
32
33
  License-File: LICENSE
33
- Requires-Dist: alembic==1.16.1
34
+ Requires-Dist: alembic==1.17.2
34
35
  Requires-Dist: anyio>=3.7.0
35
36
  Requires-Dist: apscheduler>=3.11.0
36
37
  Requires-Dist: click==8.*
37
38
  Requires-Dist: deepmerge==2.0
38
39
  Requires-Dist: deprecated>=1.2.18
39
- Requires-Dist: fastapi~=0.120.4
40
+ Requires-Dist: fastapi~=0.121.1
40
41
  Requires-Dist: fastapi-etag==0.4.0
41
42
  Requires-Dist: itsdangerous>=2.2.0
42
43
  Requires-Dist: jinja2==3.1.6
43
- Requires-Dist: more-itertools~=10.7.0
44
- Requires-Dist: nwa-stdlib~=1.10.3
44
+ Requires-Dist: more-itertools~=10.8.0
45
+ Requires-Dist: nwa-stdlib~=1.11.0
45
46
  Requires-Dist: oauth2-lib>=2.5.0
46
- Requires-Dist: orjson==3.10.18
47
+ Requires-Dist: orjson==3.11.4
47
48
  Requires-Dist: pgvector>=0.4.1
48
- Requires-Dist: prometheus-client==0.22.1
49
+ Requires-Dist: prometheus-client==0.23.1
49
50
  Requires-Dist: psycopg2-binary==2.9.11
50
51
  Requires-Dist: pydantic-forms>=1.4.0
51
- Requires-Dist: pydantic-settings~=2.9.1
52
- Requires-Dist: pydantic[email]~=2.11.7
53
- Requires-Dist: python-dateutil==2.8.2
54
- Requires-Dist: python-rapidjson>=1.18,<1.21
52
+ Requires-Dist: pydantic-settings~=2.12.0
53
+ Requires-Dist: pydantic[email]~=2.12.4
54
+ Requires-Dist: python-dateutil==2.9.0.post0
55
+ Requires-Dist: python-rapidjson>=1.22,<1.23
55
56
  Requires-Dist: pytz==2025.2
56
- Requires-Dist: redis==5.1.1
57
+ Requires-Dist: redis==5.3.1
57
58
  Requires-Dist: semver==3.0.4
58
59
  Requires-Dist: sentry-sdk[fastapi]>=2.29.1
59
60
  Requires-Dist: sqlalchemy==2.0.44
60
- Requires-Dist: sqlalchemy-utils==0.41.2
61
+ Requires-Dist: sqlalchemy-utils==0.42.0
61
62
  Requires-Dist: strawberry-graphql>=0.281.0,<0.285.0
62
63
  Requires-Dist: structlog>=25.4.0
63
64
  Requires-Dist: tabulate==0.9.0
@@ -1,4 +1,4 @@
1
- orchestrator/__init__.py,sha256=A8BOfSSiukGCfYFuZpv79RCPqojLwZU_jhE2xj4MIjU,1457
1
+ orchestrator/__init__.py,sha256=Lby3r_nYazHbsANmKYjUnfpWl5UDYWncINvMVs0RipU,1454
2
2
  orchestrator/agentic_app.py,sha256=ouiyyZiS4uS6Lox2DtbGGRnb2njJBMSHpSAGe-T5rX0,3028
3
3
  orchestrator/app.py,sha256=w8ubXaaogwjmwLM0TXqZaLkAhmaOTWzVlwiYbi5mHeE,13203
4
4
  orchestrator/exception_handlers.py,sha256=UsW3dw8q0QQlNLcV359bIotah8DYjMsj2Ts1LfX4ClY,1268
@@ -105,7 +105,7 @@ orchestrator/cli/helpers/input_helpers.py,sha256=pv5GTMuIWLzBE_bKNhn1XD_gxoqB0s1
105
105
  orchestrator/cli/helpers/print_helpers.py,sha256=b3ePg6HfBLKPYBBVr5XOA__JnFEMI5HBjbjov3CP8Po,859
106
106
  orchestrator/cli/search/__init__.py,sha256=K15_iW9ogR7xtX7qHDal4H09tmwVGnOBZWyPBLWhuzc,1274
107
107
  orchestrator/cli/search/display.py,sha256=PKy9sOTpq0WUdSfY2khLrIQ1OdAfsyl95ogF1Z6Dae0,3629
108
- orchestrator/cli/search/index_llm.py,sha256=RWPkFz5bxiznjpN1vMsSWeqcvYKB90DLL4pXQ92QJNI,2239
108
+ orchestrator/cli/search/index_llm.py,sha256=VQlLNGXro4ZbehWZIMKPHGpGE4oF1at2bkWZssimWc4,2707
109
109
  orchestrator/cli/search/resize_embedding.py,sha256=iJdM7L6Kyq4CzRjXHWLwpGRiMnKK7xZ9133C0THebBE,4847
110
110
  orchestrator/cli/search/search_explore.py,sha256=LNAn6w13Q69fpv8CHcicHAbabrctrjGvwTjjJyC0AZY,8447
111
111
  orchestrator/cli/search/speedtest.py,sha256=J_l-8WxgN3YnqmwnbRhDyVbeqtvk3d2SfIpRBOJuhvE,4840
@@ -179,16 +179,16 @@ orchestrator/graphql/mutations/customer_description.py,sha256=zm_X1yvWl4qC97_rYU
179
179
  orchestrator/graphql/mutations/start_process.py,sha256=8vLVvmBwL1ujbZJoI_8YE3VAgI-J2RTzgrTZJC8THZ4,1576
180
180
  orchestrator/graphql/resolvers/__init__.py,sha256=EEw9NO4LAryfrpkLlgsNQ9rytKd0usBDx95OURRV6sg,1031
181
181
  orchestrator/graphql/resolvers/customer.py,sha256=tq06MtMoaqFwqn3YQvSv0VmROW7QJZRJp1ykO4tUhck,934
182
- orchestrator/graphql/resolvers/helpers.py,sha256=8NeuiJD9CPus4BhRK9nDsSMhb2LhW1W7nrEpyj_J3M4,771
183
- orchestrator/graphql/resolvers/process.py,sha256=Hqs1F7-gw0yO_ioHjh2eLAyxrK2WSuL3VJ0y2ouNcLA,5010
184
- orchestrator/graphql/resolvers/product.py,sha256=uPBmYwMdau-zUqNjoDl-LDn927u3aCFW5JQ4A_it8q0,2772
185
- orchestrator/graphql/resolvers/product_block.py,sha256=Ker1CpxGab5h2BZujOHHwRUj8W4uphRr3WSpQGk2PnI,2939
186
- orchestrator/graphql/resolvers/resource_type.py,sha256=SREZXjkLYpuo4nCM8DqVeImIrZcP3xDiWr_gq4wWaxQ,2956
187
- orchestrator/graphql/resolvers/scheduled_tasks.py,sha256=QsnesRrj8ESuS9vPKG9DXYcG2Wfj9m5LWGeZgmc6hu8,1640
188
- orchestrator/graphql/resolvers/settings.py,sha256=xVYqxo-EWQ24F4hUHm9OZeN9vsqQXJzIJ1_HF4Ci9Cs,3777
189
- orchestrator/graphql/resolvers/subscription.py,sha256=57niFv-JCro_wm0peJ5Ne04F2WIPuJ-Lx2h8yd9qubA,6541
190
- orchestrator/graphql/resolvers/version.py,sha256=qgwe1msPOexeg3RHCscJ8s45vNfMhYh9ZKyCZ3MNw30,809
191
- orchestrator/graphql/resolvers/workflow.py,sha256=RKN-mI9Rg_4vVURXr943AHaftM6G3vpcRdeGgwy-uoc,2806
182
+ orchestrator/graphql/resolvers/helpers.py,sha256=RqQZrlH-YsxJ1qOnIMprrYR_K4gY2N8JFXM9PrmW15o,1215
183
+ orchestrator/graphql/resolvers/process.py,sha256=4pW80Ob8LfCqIXwTdVPE5KBpdoMGqqCVecl41HK5JEE,5034
184
+ orchestrator/graphql/resolvers/product.py,sha256=PvKPUBryO4V3NdtPzt6cSfZ3xHl8ftE8XCRDiAc4DwA,2790
185
+ orchestrator/graphql/resolvers/product_block.py,sha256=ugcrG1QyD-hjiOZdRA9EwIBkNSZbeXyXOiGaBEOlmMQ,2957
186
+ orchestrator/graphql/resolvers/resource_type.py,sha256=uELemhZPInapKhUO5oOg4fkX3v9gVij3ctD4IWg95Bo,2974
187
+ orchestrator/graphql/resolvers/scheduled_tasks.py,sha256=A_MxDp-v-gkxV9lZl_WHbeVdzWzqMDMnDwycQOC7x0c,1708
188
+ orchestrator/graphql/resolvers/settings.py,sha256=n3lLzP-aY19m9ZMDuBD0quSNstJZOMTFlk0-h0mppcU,3851
189
+ orchestrator/graphql/resolvers/subscription.py,sha256=XlA9EmSyI31djbdeCvm_uFs-8-JkU8mYCKUVI_r9euE,6694
190
+ orchestrator/graphql/resolvers/version.py,sha256=g1gLEicvXQK6iLks38hdDn8GK2k1s4cDhsa0nYXEa5E,883
191
+ orchestrator/graphql/resolvers/workflow.py,sha256=VmzE6GbkFxtfC57sP2pVg6pD1Ul10-NU3o-bxrlnHNs,2824
192
192
  orchestrator/graphql/schemas/__init__.py,sha256=dWG4DNzWq5akQ3v5tSAvT03HLxPWXa09Gx8rTz_MHmk,940
193
193
  orchestrator/graphql/schemas/customer.py,sha256=ZptVFG0qauZaoy29KDrh6k5xAnacNCTavmQrZMH8czc,147
194
194
  orchestrator/graphql/schemas/customer_description.py,sha256=fize71IMpkvk_rTzcqCYxazR5WBI652ZZ3eGvd8QqhQ,213
@@ -299,15 +299,15 @@ orchestrator/search/docs/index.md,sha256=zKzE2fbtHDfYTKaHg628wAsqCTOJ5yFUWV0ucFH
299
299
  orchestrator/search/docs/running_local_text_embedding_inference.md,sha256=OR0NVZMb8DqpgXYxlwDUrJwfRk0bYOk1-LkDMqsV6bU,1327
300
300
  orchestrator/search/filters/__init__.py,sha256=pmnHyq5SrqxS6eEiyhSIify776W9BIZ0gjOZFuYZ0nU,1335
301
301
  orchestrator/search/filters/base.py,sha256=lUr0eW0zi4oIMVUHuRD3GAQ9xEbHiFUl_EfAI6ABPVo,12456
302
- orchestrator/search/filters/date_filters.py,sha256=0a6nbUTK647_Qf4XXZMLDvBLVjF5Qqy9eJ-9SrTGaGg,3040
302
+ orchestrator/search/filters/date_filters.py,sha256=DrmOcjL3v7bh93xjC4_Q0kkZV-yanBaUetwnYTRcAGI,3045
303
303
  orchestrator/search/filters/definitions.py,sha256=k30Dp1bEr3CWMeuIcF1wPgmmF3rxI9Urx-sCaPfaE3c,4607
304
304
  orchestrator/search/filters/ltree_filters.py,sha256=1OOmM5K90NsGBQmTqyoDlphdAOGd9r2rmz1rNItm8yk,2341
305
305
  orchestrator/search/filters/numeric_filter.py,sha256=do52w5Dmb5Rt4ipfX1iEObNcdymbWHtgS2HI1Otq-JQ,2771
306
306
  orchestrator/search/indexing/__init__.py,sha256=Or78bizNPiuNOgwLGJQ0mspCF1G_gSe5C9Ap7qi0MZk,662
307
- orchestrator/search/indexing/indexer.py,sha256=2qqDe6IlKfz1exh0xLBmpPdkTqbapLnTJORZneM6tmw,15320
308
- orchestrator/search/indexing/registry.py,sha256=N4YOUhNJfY6iBwPnn76tDcZaOVYJMA2SxwWhdBE85Xs,3716
309
- orchestrator/search/indexing/tasks.py,sha256=vmS1nnprPF74yitS0xGpP1dhSDis2nekMYF0v_jduDE,2478
310
- orchestrator/search/indexing/traverse.py,sha256=NKkKSri-if1d1vwzTQlDCF0hvBdB2IbWWuMdPrQ78Jg,14330
307
+ orchestrator/search/indexing/indexer.py,sha256=4Oh-gspJrjhyecw87TK68lvGb3inVy2Sa8RlD_FHo3c,16357
308
+ orchestrator/search/indexing/registry.py,sha256=V6Q4aRXHON1gSE6wsavEIfwHwCPicSzFBS2mqNExFGs,4305
309
+ orchestrator/search/indexing/tasks.py,sha256=0p68RNwJnHSGZQjfdpyFsS2Ma5Gr2PpZROZgal_R1wI,3064
310
+ orchestrator/search/indexing/traverse.py,sha256=JLut9t4LoPCWhJ_63VgYhRKfjwyxRv-mTbQLC8mA_mU,15158
311
311
  orchestrator/search/query/__init__.py,sha256=nCjvK_n2WQdV_ACrncFXEfnvLcHtuI__J7KLlFIaQvo,2437
312
312
  orchestrator/search/query/builder.py,sha256=kgnJ93TOCm8UTL5k09nWLsG4NXAlvFFa65gbciOwZ8E,10153
313
313
  orchestrator/search/query/engine.py,sha256=TFdV_sSoSXCSDSpyhVA2S6YaJysDSW2WtPj7duAyomk,5745
@@ -315,15 +315,16 @@ orchestrator/search/query/exceptions.py,sha256=DrkNzXVbQAOi28FTHKimf_eTrXmhYwXrH
315
315
  orchestrator/search/query/export.py,sha256=_0ncVpTqN6AoQfW3WX0fWnDQX3hBz6ZGC31Beu4PVwQ,6678
316
316
  orchestrator/search/query/mixins.py,sha256=BdVDzCOFDXT6N9LI_WrbVzGrk61UNplX-UZPvD0rEV0,3019
317
317
  orchestrator/search/query/queries.py,sha256=j1uKSQgF_ifVaDJaxjs4h2z48KqGVEIKCXOoJ7Ur9Mk,3805
318
- orchestrator/search/query/results.py,sha256=8k3o7F5EESuNOdNRPsmv0klPAG9YlPisCR4iGtewiXg,8662
318
+ orchestrator/search/query/results.py,sha256=5OgAs39oncDIBdpB3NJltPr-UvLvLlxTWw9sn-lyfQA,10989
319
319
  orchestrator/search/query/state.py,sha256=fMSBJs39kZTkpDE2T4h4x0x-51GqUvzAuePg2YUbO6I,3220
320
320
  orchestrator/search/query/validation.py,sha256=m0xJ71A0Qa5hm8b71zKRjSVpPrnkG7LbqPu4lv_GboI,8260
321
321
  orchestrator/search/retrieval/__init__.py,sha256=q5G0z3nKjIHKFs1PkEG3nvTUy3Wp4kCyBtCbqUITj3A,579
322
322
  orchestrator/search/retrieval/pagination.py,sha256=kcUzq1QQk4GrZq02M4hsKwAelUo1qDeCqsXImLUK6DA,3006
323
- orchestrator/search/retrieval/retrievers/__init__.py,sha256=1bGmbae0GYRM6e1vxf0ww79NaTSmfOMe9S0pPVmh3CM,897
324
- orchestrator/search/retrieval/retrievers/base.py,sha256=Xfjo-LD_fltKy0xhBBUSsmm2v2TNygDotWQ0wTTyvVE,4155
323
+ orchestrator/search/retrieval/retrievers/__init__.py,sha256=dJlN6a0oHSquzjE5POYxrMGOXMx4Bx2khbJI-rA_qwg,971
324
+ orchestrator/search/retrieval/retrievers/base.py,sha256=esdYrkyUjwjpg-fg7BurOMe7WCTUr2cjxHqdMKDc3DI,4490
325
325
  orchestrator/search/retrieval/retrievers/fuzzy.py,sha256=PLp_ANRLzmtGQP1t9X4jt43_JLKDnOxWU2xqlexSH1U,3779
326
326
  orchestrator/search/retrieval/retrievers/hybrid.py,sha256=l-7J4qct0h28wSi0KvdFJw2lyh3jyobbrCbg0PuX-4I,11141
327
+ orchestrator/search/retrieval/retrievers/process.py,sha256=_nEEYex9iO4iBVrn6VCbvSIHf7Kb76c6id2krs-uef0,9255
327
328
  orchestrator/search/retrieval/retrievers/semantic.py,sha256=36ky_A_LNWs13IYe809qy1RPrd0Fab-G-9pf2ZDARhA,3905
328
329
  orchestrator/search/retrieval/retrievers/structured.py,sha256=13TxC52fpNGXHnPX40J2GczRYFk8LAvWn2a0HWZCd2Q,1426
329
330
  orchestrator/services/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
@@ -378,8 +379,8 @@ orchestrator/workflows/tasks/cleanup_tasks_log.py,sha256=BfWYbPXhnLAHUJ0mlODDnjZ
378
379
  orchestrator/workflows/tasks/resume_workflows.py,sha256=T3iobSJjVgiupe0rClD34kUZ7KF4pL5yK2AVeRLZog8,4313
379
380
  orchestrator/workflows/tasks/validate_product_type.py,sha256=lo2TX_MZOfcOmYFjLyD82FrJ5AAN3HOsE6BhDVFuy9Q,3210
380
381
  orchestrator/workflows/tasks/validate_products.py,sha256=GZJBoFF-WMphS7ghMs2-gqvV2iL1F0POhk0uSNt93n0,8510
381
- orchestrator/workflows/translations/en-GB.json,sha256=ST53HxkphFLTMjFHonykDBOZ7-P_KxksktZU3GbxLt0,846
382
- orchestrator_core-4.6.3rc1.dist-info/licenses/LICENSE,sha256=b-aA5OZQuuBATmLKo_mln8CQrDPPhg3ghLzjPjLn4Tg,11409
383
- orchestrator_core-4.6.3rc1.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
384
- orchestrator_core-4.6.3rc1.dist-info/METADATA,sha256=6-UwigL10OGwkHzvuZOq_ilbYhJoS_7Mmo1Y1Jpub-I,6362
385
- orchestrator_core-4.6.3rc1.dist-info/RECORD,,
382
+ orchestrator/workflows/translations/en-GB.json,sha256=Gc5gy_RghZOeSNcJIntAsz_7DsCg8n_vzoHBPXxCn_U,908
383
+ orchestrator_core-4.6.4.dist-info/licenses/LICENSE,sha256=b-aA5OZQuuBATmLKo_mln8CQrDPPhg3ghLzjPjLn4Tg,11409
384
+ orchestrator_core-4.6.4.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
385
+ orchestrator_core-4.6.4.dist-info/METADATA,sha256=wtpq1zMwlyS21ZMcodbT338EwqmKBXWV28KAcyJRStg,6416
386
+ orchestrator_core-4.6.4.dist-info/RECORD,,