dataflow-core 2.1.3__py3-none-any.whl → 2.1.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
dataflow/environment.py CHANGED
@@ -108,7 +108,7 @@ class EnvironmentManager:
108
108
  yaml_file_path=yaml_path,
109
109
  source_path=source_path,
110
110
  log_file_location=log_file_location,
111
- version=1
111
+ version="1"
112
112
  )
113
113
 
114
114
  # Update job log status if db was provided
@@ -4,10 +4,12 @@ set -e
4
4
  source_env_name=$1
5
5
  target_env_path=$2
6
6
 
7
- # Use an isolated conda package cache to avoid concurrency issues
8
- export CONDA_PKGS_DIRS=$(mktemp -d)
9
- # to delete conda package cache after script finishes
10
- trap 'rm -rf "$CONDA_PKGS_DIRS"' EXIT
7
+ # Extract just the env name (basename) from the target path
8
+ env_name=$(basename "$target_env_path")
9
+
10
+ # Set unique cache dir per environment
11
+ export CONDA_PKGS_DIRS="/dataflow/envs/cache/${env_name}"
12
+ mkdir -p "$CONDA_PKGS_DIRS"
11
13
 
12
14
  # 1. Cloning conda env
13
15
  conda create --clone ${source_env_name} --prefix ${target_env_path} --yes
@@ -17,11 +17,12 @@ if [ ! -f "$yaml_file_path" ]; then
17
17
  exit 1
18
18
  fi
19
19
 
20
- # Use an isolated conda package cache to avoid concurrency issues
21
- export CONDA_PKGS_DIRS=$(mktemp -d)
20
+ # Extract just the env name (basename) from the target path
21
+ env_name=$(basename "$conda_env_path")
22
22
 
23
- # to delete conda package cache after script finishes
24
- trap 'rm -rf "$CONDA_PKGS_DIRS"' EXIT
23
+ # Set unique cache dir per environment
24
+ export CONDA_PKGS_DIRS="/dataflow/envs/cache/${env_name}"
25
+ mkdir -p "$CONDA_PKGS_DIRS"
25
26
 
26
27
  # Create the conda environment from the YAML file
27
28
  conda env create --file "$yaml_file_path" --prefix "$conda_env_path" --yes
@@ -22,11 +22,12 @@ if [ ! -d "$conda_env_path" ]; then
22
22
  exit 1
23
23
  fi
24
24
 
25
- # Use an isolated conda package cache to avoid concurrency issues
26
- export CONDA_PKGS_DIRS=$(mktemp -d)
25
+ # Extract just the env name (basename) from the target path
26
+ env_name=$(basename "$conda_env_path")
27
27
 
28
- # to delete conda package cache after script finishes
29
- trap 'rm -rf "$CONDA_PKGS_DIRS"' EXIT
28
+ # Set unique cache dir per environment
29
+ export CONDA_PKGS_DIRS="/dataflow/envs/cache/${env_name}"
30
+ mkdir -p "$CONDA_PKGS_DIRS"
30
31
 
31
32
  # Update the conda environment using the YAML file
32
33
  conda env update --prefix "$conda_env_path" --file "$yaml_file_path" --prune
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: dataflow-core
3
- Version: 2.1.3
3
+ Version: 2.1.5
4
4
  Summary: Dataflow core package
5
5
  Author: Dataflow
6
6
  Author-email:
@@ -7,7 +7,7 @@ dataflow/configuration.py,sha256=7To6XwH1eESiYp39eqPcswXWwrdBUdPF6xN6WnazOF0,663
7
7
  dataflow/database_manager.py,sha256=tJHMuOZ9Muskrh9t4uLRlTuFU0VkHAzoHlGP5DORIC4,899
8
8
  dataflow/dataflow.py,sha256=-UYZst7EO1GgaOjlAkKu-tu7RC6XsgadGeDp1MOvZiA,7439
9
9
  dataflow/db.py,sha256=5UwE4w5Vn9RqFIhr8ARlu2haZX-PtHDLRPjmn5BG2m8,1649
10
- dataflow/environment.py,sha256=eAWYhQFM7CA3o-CA3qTWbkE8ZHZ-Jo2qvSkDQ4dKqmY,27238
10
+ dataflow/environment.py,sha256=05F-dBRyZu2mr26vuiuJU13pfgtRpA9yW69bk83HGUw,27239
11
11
  dataflow/models/__init__.py,sha256=QMLiKj8BMhfScWMm8kgHkMjwAlFeg5Cym3_AI1NvBUA,783
12
12
  dataflow/models/app_types.py,sha256=yE_ZB13lhpK7AZ7PyBwnQlf0RlIHYs_-vdMKx7_RMlY,379
13
13
  dataflow/models/blacklist_library.py,sha256=B2oi3Z8GcR_glhLAyinFk0W8c9txXvm3uOER6dY-q7I,991
@@ -29,16 +29,16 @@ dataflow/models/user.py,sha256=PT-zwZj7NWUubIj_7EY2EsjduMbI_42EyMclWMLESGk,1073
29
29
  dataflow/models/user_environment.py,sha256=yI9NutULcLiwlycuEin6ROe6o1Sjdv_sgw2MEkJFeYg,568
30
30
  dataflow/models/user_team.py,sha256=r_fmKvf6JuGgiiI9TXWjVG2QZ3WOvDrOwYWVQ3r8oWo,659
31
31
  dataflow/models/variables.py,sha256=Sinvv3zFYni5i_GrL69cVfhCh4tOOaIHiEzWYRJ-i10,1132
32
- dataflow/scripts/clone_environment.sh,sha256=dVs-NAGHtpYsk-OjoZ_gbYuZZoi3jIbEp_zXbDXEVbc,455
33
- dataflow/scripts/create_environment.sh,sha256=TLJ7FKYyhsLe0bqBy74FnpuvjFTSHiGXp1iLMcOaeJA,798
34
- dataflow/scripts/update_environment.sh,sha256=p8r2qV4blqLyC7eksHSkUDoXx_UL4Xc4NWmx8y0h_rc,962
32
+ dataflow/scripts/clone_environment.sh,sha256=xWJBw9z1W1rztrzLXYro3UtEdFuBSqNrB83y45zqFfE,487
33
+ dataflow/scripts/create_environment.sh,sha256=3FHgNplJuEZvyTsLqlCJNX9oyfXgsfqn80VZk2xtvso,828
34
+ dataflow/scripts/update_environment.sh,sha256=2dtn2xlNi6frpig-sqlGE1_IKRbbkqYOCpf_qyMKKII,992
35
35
  dataflow/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
36
36
  dataflow/utils/aws_secrets_manager.py,sha256=A_fNs9VNah9dDdl9NhqizJamYU7xr2v_GXlw9InEDFk,2380
37
37
  dataflow/utils/get_current_user.py,sha256=akjcUyTpmMdAZj9LFGSTs76hjBRjltNk9hLUqC_BdkA,1140
38
38
  dataflow/utils/json_handler.py,sha256=5_7WdypegRBDe2HSqBXyrJAdd92wsha8qRcmQvCj1TA,782
39
39
  dataflow/utils/logger.py,sha256=7BFrOq5Oiqn8P4XZbgJzMP5O07d2fpdECbbfsjrUuHw,1213
40
- dataflow_core-2.1.3.dist-info/METADATA,sha256=dRhOgxGmyMCLl9eJQ0XfExl5Fp0BYZzdJfYGgvVS3Mo,301
41
- dataflow_core-2.1.3.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
42
- dataflow_core-2.1.3.dist-info/entry_points.txt,sha256=ppj_EIbYrJJwCPg1kfdsZk5q1N-Ejfis1neYrnjhO8o,117
43
- dataflow_core-2.1.3.dist-info/top_level.txt,sha256=SZsUOpSCK9ntUy-3Tusxzf5A2e8ebwD8vouPb1dPt_8,23
44
- dataflow_core-2.1.3.dist-info/RECORD,,
40
+ dataflow_core-2.1.5.dist-info/METADATA,sha256=8UhovnGu3NJ6b4wK-xBK7QBkr17y3O96P_RGTJHf8XY,301
41
+ dataflow_core-2.1.5.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
42
+ dataflow_core-2.1.5.dist-info/entry_points.txt,sha256=ppj_EIbYrJJwCPg1kfdsZk5q1N-Ejfis1neYrnjhO8o,117
43
+ dataflow_core-2.1.5.dist-info/top_level.txt,sha256=SZsUOpSCK9ntUy-3Tusxzf5A2e8ebwD8vouPb1dPt_8,23
44
+ dataflow_core-2.1.5.dist-info/RECORD,,