dbos 1.1.0a4__tar.gz → 1.2.0a2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (106) hide show
  1. {dbos-1.1.0a4 → dbos-1.2.0a2}/PKG-INFO +1 -1
  2. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_admin_server.py +24 -4
  3. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_app_db.py +0 -15
  4. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_core.py +0 -1
  5. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_dbos.py +22 -0
  6. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_event_loop.py +10 -7
  7. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_sys_db.py +0 -12
  8. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_workflow_commands.py +1 -10
  9. {dbos-1.1.0a4 → dbos-1.2.0a2}/pyproject.toml +1 -1
  10. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_admin_server.py +28 -1
  11. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_async.py +42 -1
  12. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_workflow_management.py +0 -11
  13. {dbos-1.1.0a4 → dbos-1.2.0a2}/LICENSE +0 -0
  14. {dbos-1.1.0a4 → dbos-1.2.0a2}/README.md +0 -0
  15. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/__init__.py +0 -0
  16. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/__main__.py +0 -0
  17. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_classproperty.py +0 -0
  18. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_client.py +0 -0
  19. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_conductor/conductor.py +0 -0
  20. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_conductor/protocol.py +0 -0
  21. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_context.py +0 -0
  22. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_croniter.py +0 -0
  23. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_dbos_config.py +0 -0
  24. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_debug.py +0 -0
  25. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_docker_pg_helper.py +0 -0
  26. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_error.py +0 -0
  27. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_fastapi.py +0 -0
  28. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_flask.py +0 -0
  29. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_kafka.py +0 -0
  30. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_kafka_message.py +0 -0
  31. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_logger.py +0 -0
  32. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/env.py +0 -0
  33. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/script.py.mako +0 -0
  34. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/04ca4f231047_workflow_queues_executor_id.py +0 -0
  35. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/27ac6900c6ad_add_queue_dedup.py +0 -0
  36. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/50f3227f0b4b_fix_job_queue.py +0 -0
  37. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/5c361fc04708_added_system_tables.py +0 -0
  38. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/83f3732ae8e7_workflow_timeout.py +0 -0
  39. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/933e86bdac6a_add_queue_priority.py +0 -0
  40. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/a3b18ad34abe_added_triggers.py +0 -0
  41. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/d76646551a6b_job_queue_limiter.py +0 -0
  42. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/d76646551a6c_workflow_queue.py +0 -0
  43. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/eab0cc1d9a14_job_queue.py +0 -0
  44. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_migrations/versions/f4b9b32ba814_functionname_childid_op_outputs.py +0 -0
  45. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_outcome.py +0 -0
  46. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_queue.py +0 -0
  47. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_recovery.py +0 -0
  48. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_registrations.py +0 -0
  49. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_roles.py +0 -0
  50. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_scheduler.py +0 -0
  51. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_schemas/__init__.py +0 -0
  52. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_schemas/application_database.py +0 -0
  53. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_schemas/system_database.py +0 -0
  54. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_serialization.py +0 -0
  55. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/README.md +0 -0
  56. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/__package/__init__.py +0 -0
  57. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/__package/main.py.dbos +0 -0
  58. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/__package/schema.py +0 -0
  59. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/alembic.ini +0 -0
  60. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/dbos-config.yaml.dbos +0 -0
  61. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/migrations/env.py.dbos +0 -0
  62. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/migrations/script.py.mako +0 -0
  63. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/migrations/versions/2024_07_31_180642_init.py +0 -0
  64. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_templates/dbos-db-starter/start_postgres_docker.py +0 -0
  65. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_tracer.py +0 -0
  66. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/_utils.py +0 -0
  67. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/cli/_github_init.py +0 -0
  68. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/cli/_template_init.py +0 -0
  69. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/cli/cli.py +0 -0
  70. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/dbos-config.schema.json +0 -0
  71. {dbos-1.1.0a4 → dbos-1.2.0a2}/dbos/py.typed +0 -0
  72. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/__init__.py +0 -0
  73. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/atexit_no_ctor.py +0 -0
  74. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/atexit_no_launch.py +0 -0
  75. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/classdefs.py +0 -0
  76. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/client_collateral.py +0 -0
  77. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/client_worker.py +0 -0
  78. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/conftest.py +0 -0
  79. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/dupname_classdefs1.py +0 -0
  80. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/dupname_classdefsa.py +0 -0
  81. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/more_classdefs.py +0 -0
  82. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/queuedworkflow.py +0 -0
  83. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_classdecorators.py +0 -0
  84. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_cli.py +0 -0
  85. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_client.py +0 -0
  86. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_concurrency.py +0 -0
  87. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_config.py +0 -0
  88. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_croniter.py +0 -0
  89. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_dbos.py +0 -0
  90. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_debug.py +0 -0
  91. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_docker_secrets.py +0 -0
  92. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_failures.py +0 -0
  93. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_fastapi.py +0 -0
  94. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_fastapi_roles.py +0 -0
  95. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_flask.py +0 -0
  96. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_kafka.py +0 -0
  97. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_outcome.py +0 -0
  98. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_package.py +0 -0
  99. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_queue.py +0 -0
  100. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_scheduler.py +0 -0
  101. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_schema_migration.py +0 -0
  102. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_singleton.py +0 -0
  103. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_spans.py +0 -0
  104. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_sqlalchemy.py +0 -0
  105. {dbos-1.1.0a4 → dbos-1.2.0a2}/tests/test_workflow_introspection.py +0 -0
  106. {dbos-1.1.0a4 → dbos-1.2.0a2}/version/__init__.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: dbos
3
- Version: 1.1.0a4
3
+ Version: 1.2.0a2
4
4
  Summary: Ultra-lightweight durable execution in Python
5
5
  Author-Email: "DBOS, Inc." <contact@dbos.dev>
6
6
  License: MIT
@@ -5,8 +5,9 @@ import re
5
5
  import threading
6
6
  from functools import partial
7
7
  from http.server import BaseHTTPRequestHandler, ThreadingHTTPServer
8
- from typing import TYPE_CHECKING, Any, List, TypedDict
8
+ from typing import TYPE_CHECKING, Any, List, Optional, TypedDict
9
9
 
10
+ from ._context import SetWorkflowID
10
11
  from ._error import DBOSException
11
12
  from ._logger import dbos_logger
12
13
  from ._recovery import recover_pending_workflows
@@ -141,7 +142,11 @@ class AdminRequestHandler(BaseHTTPRequestHandler):
141
142
  try:
142
143
  data = json.loads(post_data.decode("utf-8"))
143
144
  start_step: int = data.get("start_step", 1)
144
- self._handle_fork(workflow_id, start_step)
145
+ new_workflow_id: Optional[str] = data.get("new_workflow_id")
146
+ application_version: Optional[str] = data.get("application_version")
147
+ self._handle_fork(
148
+ workflow_id, start_step, new_workflow_id, application_version
149
+ )
145
150
  except (json.JSONDecodeError, AttributeError) as e:
146
151
  self.send_response(500)
147
152
  self.send_header("Content-Type", "application/json")
@@ -191,9 +196,24 @@ class AdminRequestHandler(BaseHTTPRequestHandler):
191
196
  self.end_headers()
192
197
  self.wfile.write(response_body)
193
198
 
194
- def _handle_fork(self, workflow_id: str, start_step: int) -> None:
199
+ def _handle_fork(
200
+ self,
201
+ workflow_id: str,
202
+ start_step: int,
203
+ new_workflow_id: Optional[str],
204
+ application_version: Optional[str],
205
+ ) -> None:
195
206
  try:
196
- handle = self.dbos.fork_workflow(workflow_id, start_step)
207
+ print(f"Forking workflow {workflow_id} from step {start_step}")
208
+ if new_workflow_id is not None:
209
+ with SetWorkflowID(new_workflow_id):
210
+ handle = self.dbos.fork_workflow(
211
+ workflow_id, start_step, application_version=application_version
212
+ )
213
+ else:
214
+ handle = self.dbos.fork_workflow(
215
+ workflow_id, start_step, application_version=application_version
216
+ )
197
217
  response_body = json.dumps(
198
218
  {
199
219
  "workflow_id": handle.workflow_id,
@@ -216,21 +216,6 @@ class ApplicationDatabase:
216
216
  for row in rows
217
217
  ]
218
218
 
219
- def get_max_function_id(self, workflow_uuid: str) -> Optional[int]:
220
- with self.engine.begin() as conn:
221
- max_function_id_row = conn.execute(
222
- sa.select(
223
- sa.func.max(ApplicationSchema.transaction_outputs.c.function_id)
224
- ).where(
225
- ApplicationSchema.transaction_outputs.c.workflow_uuid
226
- == workflow_uuid
227
- )
228
- ).fetchone()
229
-
230
- max_function_id = max_function_id_row[0] if max_function_id_row else None
231
-
232
- return max_function_id
233
-
234
219
  def clone_workflow_transactions(
235
220
  self, src_workflow_id: str, forked_workflow_id: str, start_step: int
236
221
  ) -> None:
@@ -602,7 +602,6 @@ async def start_workflow_async(
602
602
  *args: P.args,
603
603
  **kwargs: P.kwargs,
604
604
  ) -> "WorkflowHandleAsync[R]":
605
-
606
605
  # If the function has a class, add the class object as its first argument
607
606
  fself: Optional[object] = None
608
607
  if hasattr(func, "__self__"):
@@ -299,6 +299,7 @@ class DBOS:
299
299
 
300
300
  self._launched: bool = False
301
301
  self._debug_mode: bool = False
302
+ self._configured_threadpool: bool = False
302
303
  self._sys_db_field: Optional[SystemDatabase] = None
303
304
  self._app_db_field: Optional[ApplicationDatabase] = None
304
305
  self._registry: DBOSRegistry = _get_or_create_dbos_registry()
@@ -719,6 +720,7 @@ class DBOS:
719
720
  **kwargs: P.kwargs,
720
721
  ) -> WorkflowHandleAsync[R]:
721
722
  """Invoke a workflow function on the event loop, returning a handle to the ongoing execution."""
723
+ await cls._configure_asyncio_thread_pool()
722
724
  return await start_workflow_async(
723
725
  _get_dbos_instance(), func, None, True, *args, **kwargs
724
726
  )
@@ -736,6 +738,7 @@ class DBOS:
736
738
  async def get_workflow_status_async(
737
739
  cls, workflow_id: str
738
740
  ) -> Optional[WorkflowStatus]:
741
+ await cls._configure_asyncio_thread_pool()
739
742
  """Return the status of a workflow execution."""
740
743
  return await asyncio.to_thread(cls.get_workflow_status, workflow_id)
741
744
 
@@ -757,6 +760,7 @@ class DBOS:
757
760
  ) -> WorkflowHandleAsync[R]:
758
761
  """Return a `WorkflowHandle` for a workflow execution."""
759
762
  dbos = _get_dbos_instance()
763
+ await cls._configure_asyncio_thread_pool()
760
764
  if existing_workflow:
761
765
  stat = await dbos.get_workflow_status_async(workflow_id)
762
766
  if stat is None:
@@ -775,6 +779,7 @@ class DBOS:
775
779
  cls, destination_id: str, message: Any, topic: Optional[str] = None
776
780
  ) -> None:
777
781
  """Send a message to a workflow execution."""
782
+ await cls._configure_asyncio_thread_pool()
778
783
  await asyncio.to_thread(lambda: DBOS.send(destination_id, message, topic))
779
784
 
780
785
  @classmethod
@@ -797,6 +802,7 @@ class DBOS:
797
802
  This function is to be called from within a workflow.
798
803
  `recv_async` will return the message sent on `topic`, asyncronously waiting if necessary.
799
804
  """
805
+ await cls._configure_asyncio_thread_pool()
800
806
  return await asyncio.to_thread(lambda: DBOS.recv(topic, timeout_seconds))
801
807
 
802
808
  @classmethod
@@ -835,6 +841,7 @@ class DBOS:
835
841
  It is important to use `DBOS.sleep` or `DBOS.sleep_async` (as opposed to any other sleep) within workflows,
836
842
  as the DBOS sleep methods are durable and completed sleeps will be skipped during recovery.
837
843
  """
844
+ await cls._configure_asyncio_thread_pool()
838
845
  await asyncio.to_thread(lambda: DBOS.sleep(seconds))
839
846
 
840
847
  @classmethod
@@ -869,6 +876,7 @@ class DBOS:
869
876
  value(Any): A serializable value to associate with the key
870
877
 
871
878
  """
879
+ await cls._configure_asyncio_thread_pool()
872
880
  await asyncio.to_thread(lambda: DBOS.set_event(key, value))
873
881
 
874
882
  @classmethod
@@ -901,6 +909,7 @@ class DBOS:
901
909
  timeout_seconds(float): The amount of time to wait, in case `set_event` has not yet been called byt the workflow
902
910
 
903
911
  """
912
+ await cls._configure_asyncio_thread_pool()
904
913
  return await asyncio.to_thread(
905
914
  lambda: DBOS.get_event(workflow_id, key, timeout_seconds)
906
915
  )
@@ -929,6 +938,19 @@ class DBOS:
929
938
  fn, "DBOS.cancelWorkflow"
930
939
  )
931
940
 
941
+ @classmethod
942
+ async def _configure_asyncio_thread_pool(cls) -> None:
943
+ """
944
+ Configure the thread pool for asyncio.to_thread.
945
+
946
+ This function is called before the first call to asyncio.to_thread.
947
+ """
948
+ if _get_dbos_instance()._configured_threadpool:
949
+ return
950
+ loop = asyncio.get_running_loop()
951
+ loop.set_default_executor(_get_dbos_instance()._executor)
952
+ _get_dbos_instance()._configured_threadpool = True
953
+
932
954
  @classmethod
933
955
  def resume_workflow(cls, workflow_id: str) -> WorkflowHandle[Any]:
934
956
  """Resume a workflow by ID."""
@@ -1,5 +1,6 @@
1
1
  import asyncio
2
2
  import threading
3
+ from concurrent.futures import ThreadPoolExecutor
3
4
  from typing import Any, Coroutine, Optional, TypeVar
4
5
 
5
6
 
@@ -33,15 +34,17 @@ class BackgroundEventLoop:
33
34
 
34
35
  def _run_event_loop(self) -> None:
35
36
  self._loop = asyncio.new_event_loop()
36
- asyncio.set_event_loop(self._loop)
37
+ with ThreadPoolExecutor(max_workers=64) as thread_pool:
38
+ self._loop.set_default_executor(thread_pool)
39
+ asyncio.set_event_loop(self._loop)
37
40
 
38
- self._running = True
39
- self._ready.set() # Signal that the loop is ready
41
+ self._running = True
42
+ self._ready.set() # Signal that the loop is ready
40
43
 
41
- try:
42
- self._loop.run_forever()
43
- finally:
44
- self._loop.close()
44
+ try:
45
+ self._loop.run_forever()
46
+ finally:
47
+ self._loop.close()
45
48
 
46
49
  async def _shutdown(self) -> None:
47
50
  if self._loop is None:
@@ -601,18 +601,6 @@ class SystemDatabase:
601
601
  )
602
602
  )
603
603
 
604
- def get_max_function_id(self, workflow_uuid: str) -> Optional[int]:
605
- with self.engine.begin() as conn:
606
- max_function_id_row = conn.execute(
607
- sa.select(
608
- sa.func.max(SystemSchema.operation_outputs.c.function_id)
609
- ).where(SystemSchema.operation_outputs.c.workflow_uuid == workflow_uuid)
610
- ).fetchone()
611
-
612
- max_function_id = max_function_id_row[0] if max_function_id_row else None
613
-
614
- return max_function_id
615
-
616
604
  def fork_workflow(
617
605
  self,
618
606
  original_workflow_id: str,
@@ -103,16 +103,7 @@ def fork_workflow(
103
103
  *,
104
104
  application_version: Optional[str],
105
105
  ) -> str:
106
- def get_max_function_id(workflow_uuid: str) -> int:
107
- max_transactions = app_db.get_max_function_id(workflow_uuid) or 0
108
- max_operations = sys_db.get_max_function_id(workflow_uuid) or 0
109
- return max(max_transactions, max_operations)
110
-
111
- max_function_id = get_max_function_id(workflow_id)
112
- if max_function_id > 0 and start_step > max_function_id:
113
- raise DBOSException(
114
- f"Cannot fork workflow {workflow_id} from step {start_step}. The workflow has {max_function_id} steps."
115
- )
106
+
116
107
  ctx = get_local_dbos_context()
117
108
  if ctx is not None and len(ctx.id_assigned_for_next_workflow) > 0:
118
109
  forked_workflow_id = ctx.id_assigned_for_next_workflow
@@ -27,7 +27,7 @@ dependencies = [
27
27
  ]
28
28
  requires-python = ">=3.9"
29
29
  readme = "README.md"
30
- version = "1.1.0a4"
30
+ version = "1.2.0a2"
31
31
 
32
32
  [project.license]
33
33
  text = "MIT"
@@ -11,7 +11,14 @@ import sqlalchemy as sa
11
11
  from requests.exceptions import ConnectionError
12
12
 
13
13
  # Public API
14
- from dbos import DBOS, DBOSConfig, Queue, SetWorkflowID, _workflow_commands
14
+ from dbos import (
15
+ DBOS,
16
+ DBOSConfig,
17
+ Queue,
18
+ SetWorkflowID,
19
+ WorkflowHandle,
20
+ _workflow_commands,
21
+ )
15
22
  from dbos._error import DBOSWorkflowCancelledError
16
23
  from dbos._schemas.system_database import SystemSchema
17
24
  from dbos._sys_db import SystemDatabase, WorkflowStatusString
@@ -425,4 +432,24 @@ def test_admin_workflow_fork(dbos: DBOS, sys_db: SystemDatabase) -> None:
425
432
  time.sleep(1)
426
433
  count += 1
427
434
 
435
+ # test for new_workflow_id and app version
436
+
437
+ new_version = "my_new_version"
438
+ GlobalParams.app_version = new_version
439
+
440
+ response = requests.post(
441
+ f"http://localhost:3001/workflows/{wfUuid}/fork",
442
+ json={"new_workflow_id": "123456", "application_version": new_version},
443
+ timeout=5,
444
+ )
445
+ assert response.status_code == 200
446
+
447
+ new_workflow_id = response.json().get("workflow_id")
448
+ assert new_workflow_id == "123456", "Expected new workflow ID is not 123456"
449
+
450
+ handle: WorkflowHandle[None] = dbos.retrieve_workflow(new_workflow_id)
451
+ assert (
452
+ handle.get_status().app_version == new_version
453
+ ), f"Expected application version to be {new_version}, but got {handle.get_status().app_version}"
454
+
428
455
  assert worked, "Workflow did not finish successfully"
@@ -1,7 +1,7 @@
1
1
  import asyncio
2
2
  import time
3
3
  import uuid
4
- from typing import Optional
4
+ from typing import List, Optional
5
5
 
6
6
  import pytest
7
7
  import sqlalchemy as sa
@@ -523,3 +523,44 @@ async def test_workflow_timeout_async(dbos: DBOS) -> None:
523
523
  with pytest.raises(Exception) as exc_info:
524
524
  await (await DBOS.retrieve_workflow_async(direct_child)).get_result()
525
525
  assert "was cancelled" in str(exc_info.value)
526
+
527
+
528
+ @pytest.mark.asyncio
529
+ async def test_max_parallel_workflows(dbos: DBOS) -> None:
530
+ queue = Queue("parallel_queue")
531
+
532
+ @DBOS.workflow()
533
+ async def test_workflow(i: int) -> int:
534
+ await DBOS.sleep_async(5)
535
+ return i
536
+
537
+ begin_time = time.time()
538
+
539
+ tasks: List[WorkflowHandleAsync[int]] = []
540
+ for i in range(50):
541
+ tasks.append(await DBOS.start_workflow_async(test_workflow, i))
542
+
543
+ # Wait for all tasks to complete
544
+ for i in range(50):
545
+ assert (await tasks[i].get_result()) == i, f"Task {i} should return {i}"
546
+
547
+ end_time = time.time()
548
+ assert (
549
+ end_time - begin_time < 10
550
+ ), "All tasks should complete in less than 10 seconds"
551
+
552
+ # Test enqueues
553
+ begin_time = time.time()
554
+ tasks = []
555
+
556
+ for i in range(50):
557
+ tasks.append(await queue.enqueue_async(test_workflow, i))
558
+
559
+ # Wait for all tasks to complete
560
+ for i in range(50):
561
+ assert (await tasks[i].get_result()) == i, f"Task {i} should return {i}"
562
+
563
+ end_time = time.time()
564
+ assert (
565
+ end_time - begin_time < 10
566
+ ), "All enqueued tasks should complete in less than 10 seconds"
@@ -482,17 +482,6 @@ def test_restart_fromsteps_steps_tr(
482
482
  assert stepFourCount == 2
483
483
  assert trFiveCount == 3
484
484
 
485
- # invalid step
486
- try:
487
- forked_handle = DBOS.fork_workflow(wfid, 7)
488
- assert forked_handle.workflow_id != wfid
489
- forked_handle.get_result()
490
- except Exception as e:
491
- print(f"Exception: {e}")
492
- assert isinstance(e, DBOSException)
493
- assert "Cannot fork workflow" in str(e)
494
- assert trOneCount == 1
495
-
496
485
  # invalid < 1 will default to 1
497
486
  forked_handle = DBOS.fork_workflow(wfid, -1)
498
487
  assert forked_handle.workflow_id != wfid
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes