pilot.linkstec 0.0.90__py3-none-any.whl → 0.0.92__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
File without changes
@@ -0,0 +1,75 @@
1
+ import requests
2
+
3
+ from pilot.config.config_reader import get_config
4
+ from pilot.logging.logger import get_logger
5
+
6
+
7
+ class AIClient:
8
+ def __init__(self):
9
+ self.logger = get_logger(__name__)
10
+ self.config_dto = get_config()
11
+ self.headers = {"Content-Type": "application/json;charset=utf-8"}
12
+
13
+ def call(self, user_prompt: str, system_prompt: str = "") -> str:
14
+ messages = []
15
+ if system_prompt:
16
+ messages.append({"role": "system", "content": system_prompt})
17
+ messages.append({"role": "user", "content": user_prompt})
18
+
19
+ request_data = self._build_request_payload(messages)
20
+
21
+ response_data = self._send_post_request(self.api_url, self.headers, request_data)
22
+ if not isinstance(response_data, dict):
23
+ self.logger.error("無効なAPI応答またはリクエスト失敗")
24
+ return ""
25
+
26
+ result_text = self._extract_response_content(response_data)
27
+ return result_text
28
+
29
+ def _build_request_payload(self, messages: list[dict]) -> dict:
30
+ raise NotImplementedError("サブクラスでリクエストペイロードの構築を実装してください")
31
+
32
+ def _send_post_request(self, url: str, headers: dict, data: dict) -> dict or str:
33
+ try:
34
+ response = requests.post(url, headers=headers, json=data)
35
+ except Exception as e:
36
+ self.logger.error(f"リクエスト失敗: {e}")
37
+ return ""
38
+ if response.status_code != 200:
39
+ self.logger.error(f"ステータスコード {response.status_code}: {response.text}")
40
+ return ""
41
+ try:
42
+ return response.json()
43
+ except Exception as e:
44
+ self.logger.error(f"JSON解析失敗: {e}")
45
+ return ""
46
+
47
+ def _extract_response_content(self, response: dict) -> str:
48
+ raise NotImplementedError("サブクラスでレスポンスの解析を実装してください")
49
+
50
+
51
+ class LMStudioClient(AIClient):
52
+ def __init__(self):
53
+ super().__init__()
54
+ self.api_url = self.config_dto.lm_studio_api_url
55
+ self.model_name = self.config_dto.lm_studio_model_name
56
+
57
+ def _build_request_payload(self, messages: list[dict]) -> dict:
58
+ payload = {
59
+ "model": self.model_name,
60
+ "stream": False,
61
+ "temperature": 0.8,
62
+ "max_tokens": 15000,
63
+ "messages": messages,
64
+ }
65
+ return payload
66
+
67
+ def _extract_response_content(self, response: dict) -> str:
68
+ if not isinstance(response, dict):
69
+ return str(response)
70
+ if "usage" in response:
71
+ self.logger.debug(f"使用状況: {response['usage']}")
72
+ choices = response.get("choices", [])
73
+ if choices:
74
+ return choices[0].get("message", {}).get("content") or str(response)
75
+ return str(response)
@@ -1,39 +1,54 @@
1
1
  import configparser
2
- import os
3
2
  import inspect
3
+ import os
4
4
  from dataclasses import dataclass
5
- from typing import List
5
+ from typing import Optional
6
+
6
7
 
7
8
  @dataclass
8
9
  class ConfigDTO:
9
- work_space: str
10
- threads: int
11
10
  project: str
11
+ log_level: str
12
+ threads: int
13
+ lm_studio_api_url: str
14
+ lm_studio_model_name: str
15
+ work_space: str
16
+ copy_path: str
17
+ json_file_path:str
12
18
  steps: list[str]
13
- skipsteps: list[str]
14
- runsteps: list[str]
15
- multisteps: list[str]
19
+
16
20
 
17
21
  class ConfigReader:
18
- def __init__(self, filename = None):
19
- filepath = None
20
- if filename is None:
21
- filepath = self.find_config_path()
22
-
23
- if filename is not None:
24
- cwd = os.getcwd()
25
- filepath = os.path.join(cwd, 'config', filename)
26
- if not os.path.exists(filepath):
27
- raise FileNotFoundError(f"Configuration file not found: {filepath}")
28
-
29
- self.config = configparser.ConfigParser()
30
- self.config.optionxform = str
31
-
32
- with open(filepath, 'r', encoding='utf-8') as f:
33
- content = f.read()
34
- if not content.lstrip().startswith('['):
35
- content = '[DEFAULT]\n' + content
36
- self.config.read_string(content)
22
+ _instance = None
23
+ _loaded = False
24
+
25
+ def __new__(cls, filepath=None):
26
+ if cls._instance is None:
27
+ cls._instance = super().__new__(cls)
28
+ return cls._instance
29
+
30
+ def __init__(self, filepath=None):
31
+ if self._loaded:
32
+ return
33
+
34
+ try:
35
+ if filepath is None:
36
+ filepath = self.find_config_path()
37
+ if not os.path.exists(filepath):
38
+ raise FileNotFoundError(f"設定ファイルが見つかりません: {filepath}")
39
+
40
+ self.config = configparser.ConfigParser()
41
+ self.config.optionxform = str
42
+ with open(filepath, 'r', encoding='utf-8') as f:
43
+ content = f.read()
44
+ if not content.lstrip().startswith('['):
45
+ content = '[DEFAULT]\n' + content
46
+ self.config.read_string(content)
47
+
48
+ self._loaded = True
49
+ except Exception as e:
50
+ print(f"設定ファイル読み込みエラー: {e}")
51
+ raise
37
52
 
38
53
  @classmethod
39
54
  def find_config_path(cls):
@@ -52,10 +67,11 @@ class ConfigReader:
52
67
 
53
68
  base_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', '..'))
54
69
  fallback_path = os.path.join(base_dir, 'config', 'control.properties')
70
+
55
71
  if os.path.exists(fallback_path):
56
72
  return fallback_path
57
73
 
58
- raise FileNotFoundError("control.properties not found in expected locations")
74
+ raise FileNotFoundError("control.properties が期待される場所に見つかりません")
59
75
 
60
76
  def get(self, section, option, fallback=None, cast_type=str):
61
77
  try:
@@ -71,25 +87,47 @@ class ConfigReader:
71
87
  return fallback
72
88
 
73
89
  def get_dto(self) -> ConfigDTO:
74
- input_path = self.get('DEFAULT', 'input_path', fallback='.')
75
- work_space = self.get('DEFAULT', 'work_space', fallback='.')
76
- threads = int(self.get('DEFAULT', 'threads', fallback=1))
77
90
  project = self.get('DEFAULT', 'project', fallback='')
91
+ log_level = self.get('DEFAULT', 'log_level', fallback='INFO')
92
+ threads = self.get('DEFAULT', 'threads', fallback=1, cast_type=int)
93
+ lm_studio_api_url = self.get('DEFAULT', 'lm_studio_api_url', fallback='.')
94
+ lm_studio_model_name = self.get('DEFAULT', 'lm_studio_model_name', fallback='.')
95
+ work_space = self.get('DEFAULT', 'work_space', fallback='.')
96
+ copy_path = self.get('DEFAULT', 'copy_file_path', fallback='.')
97
+ json_file_path = self.get('DEFAULT', 'json_file_path', fallback='.')
78
98
  steps_str = self.get('DEFAULT', 'steps', fallback='')
99
+
79
100
  steps = [s.strip() for s in steps_str.split(',')] if steps_str else []
80
- skipsteps_str = self.get('DEFAULT', 'skipsteps', fallback='')
81
- skipsteps = [s.strip() for s in skipsteps_str.split(',')] if skipsteps_str else []
82
- runsteps_str = self.get('DEFAULT', 'runsteps', fallback='')
83
- runsteps = [s.strip() for s in runsteps_str.split(',')] if runsteps_str else []
84
- multisteps_str = self.get('DEFAULT', 'multisteps', fallback='')
85
- multisteps = [s.strip() for s in multisteps_str.split(',')] if multisteps_str else []
86
101
 
87
102
  return ConfigDTO(
88
- work_space=work_space,
89
- threads=threads,
90
103
  project=project,
91
- steps=steps,
92
- skipsteps=skipsteps,
93
- runsteps=runsteps,
94
- multisteps=multisteps
95
- )
104
+ log_level=log_level,
105
+ threads=threads,
106
+ lm_studio_api_url=lm_studio_api_url,
107
+ lm_studio_model_name=lm_studio_model_name,
108
+ work_space=work_space,
109
+ copy_path=copy_path,
110
+ json_file_path=json_file_path,
111
+ steps=steps
112
+ )
113
+
114
+
115
+ # 全局設定管理用の変数
116
+ _global_config: Optional[ConfigDTO] = None
117
+
118
+
119
+ def init_config(filepath=None):
120
+ global _global_config
121
+ try:
122
+ config_reader = ConfigReader(filepath)
123
+ _global_config = config_reader.get_dto()
124
+ except Exception as e:
125
+ print(f"設定初期化エラー: {e}")
126
+ raise
127
+
128
+
129
+ def get_config() -> ConfigDTO:
130
+ global _global_config
131
+ if _global_config is None:
132
+ raise RuntimeError("設定が初期化されていません。init_config() を最初に呼び出してください。")
133
+ return _global_config
@@ -1,19 +1,12 @@
1
1
  import threading
2
2
  from typing import Dict, Any, Optional
3
-
4
- import tiktoken
5
- from vertexai.generative_models import GenerativeModel, ChatSession
6
- import os
3
+ import requests
7
4
 
8
5
  class VertexAISingleton:
9
6
  _instance: Optional['VertexAISingleton'] = None
10
-
11
-
12
7
  _lock = threading.Lock()
13
- _tokenizer_cache = {}
14
- encoding = None
15
8
 
16
- def __new__(cls, model_name: str = "gemini-2.5-pro"):
9
+ def __new__(cls, model_name: str = "openai/gpt-oss-20b"):
17
10
  if cls._instance is None:
18
11
  with cls._lock:
19
12
  if cls._instance is None:
@@ -21,24 +14,47 @@ class VertexAISingleton:
21
14
  cls._instance._initialized = False
22
15
  return cls._instance
23
16
 
24
- def __init__(self, model_name: str = "gemini-2.5-pro"):
17
+ def __init__(self, model_name: str = "openai/gpt-oss-20b"):
25
18
  if not self._initialized:
26
19
  with self._lock:
27
20
  if not self._initialized:
28
- self.model = GenerativeModel(model_name)
29
- self.encoding = tiktoken.get_encoding("cl100k_base")
21
+ self.model_name = model_name
22
+ self.base_url = "http://127.0.0.1:3000/v1"
23
+
24
+ #self.encoding = tiktoken.get_encoding("cl100k_base")
25
+
26
+ self._session = requests.Session()
27
+
30
28
  self._initialized = True
31
29
 
32
30
  def generate_content(self, prompt: str) -> Dict[str, Any]:
33
31
  """複数スレッドから安全に呼び出し可能"""
34
32
  try:
35
- response = self.model.generate_content(prompt)
33
+ payload = {
34
+ "model": self.model_name,
35
+ "messages": [
36
+ {"role": "user", "content": prompt}
37
+ ],
38
+ "temperature": 0.7
39
+ }
40
+
41
+ resp = self._session.post(
42
+ f"{self.base_url}/chat/completions",
43
+ json=payload,
44
+ timeout=600
45
+ )
46
+ resp.raise_for_status()
47
+ data = resp.json()
48
+
49
+ content = data["choices"][0]["message"]["content"]
50
+
36
51
  return {
37
52
  "prompt": prompt,
38
- "response": self._remove_code_fence(response.text),
53
+ "response": self._remove_code_fence(content),
39
54
  "success": True,
40
55
  "error": None
41
56
  }
57
+
42
58
  except Exception as e:
43
59
  return {
44
60
  "prompt": prompt,
@@ -47,18 +63,20 @@ class VertexAISingleton:
47
63
  "error": str(e)
48
64
  }
49
65
 
50
- def start_chat(self) -> ChatSession:
51
- """新しいチャットセッションを開始"""
52
- return self.model.start_chat()
66
+ def start_chat(self):
67
+ """
68
+ VertexAI の ChatSession と完全互換は不可能だが、
69
+ 既存コードを壊さないために「退化実装」を提供
70
+ """
71
+ return _LMStudioChatSession(self)
53
72
 
54
73
  def count_tokens(self, text: str) -> int:
55
- """与えられたテキストのトークン数を返す(bert-base-uncasedのみ使用)"""
56
- try:
57
- tokens = self.encoding.encode(text)
58
- return len(tokens)
59
- except Exception as e:
60
- print(f"トークン計算失敗: {e}")
61
- return 0
74
+ return 1
75
+ #try:
76
+ # return len(self.encoding.encode(text))
77
+ #except Exception as e:
78
+ # print(f"トークン計算失敗: {e}")
79
+ # return 0
62
80
 
63
81
  def _remove_code_fence(self, text: str) -> str:
64
82
  lines = text.splitlines()
@@ -69,6 +87,38 @@ class VertexAISingleton:
69
87
  return "\n".join(lines)
70
88
 
71
89
  @classmethod
72
- def get_instance(cls, model_name: str = "gemini-2.5-pro") -> 'VertexAISingleton':
73
- """インスタンスを取得"""
74
- return cls(model_name)
90
+ def get_instance(cls, model_name: str = "openai/gpt-oss-20b") -> 'VertexAISingleton':
91
+ return cls(model_name)
92
+
93
+ class _LMStudioChatSession:
94
+ """
95
+ VertexAI ChatSession の「最低限互換」
96
+ """
97
+ def __init__(self, client: VertexAISingleton):
98
+ self._client = client
99
+ self._messages = []
100
+
101
+ def send_message(self, message: str):
102
+ self._messages.append({"role": "user", "content": message})
103
+
104
+ payload = {
105
+ "model": self._client.model_name,
106
+ "messages": self._messages
107
+ }
108
+
109
+ resp = self._client._session.post(
110
+ f"{self._client.base_url}/chat/completions",
111
+ json=payload,
112
+ timeout=60
113
+ )
114
+ resp.raise_for_status()
115
+ data = resp.json()
116
+
117
+ reply = data["choices"][0]["message"]["content"]
118
+ self._messages.append({"role": "assistant", "content": reply})
119
+
120
+ class _Resp:
121
+ def __init__(self, text):
122
+ self.text = text
123
+
124
+ return _Resp(reply)
File without changes
File without changes
@@ -0,0 +1,16 @@
1
+ import threading
2
+
3
+ from pilot.job.impl.base_job import BaseJob
4
+
5
+ from pilot.conver.converfileEncodding import nkf_convert
6
+
7
+
8
+ class EncodingTransformerJob(BaseJob):
9
+ _begin_file_lock = threading.Lock()
10
+ def run(self):
11
+ with self._begin_file_lock:
12
+ if not self.change_current_trg_to_begin():
13
+ return
14
+ nkf_args = ['-w', '--overwrite']
15
+ nkf_convert(self.file_path, nkf_args)
16
+ super().run()
@@ -0,0 +1,27 @@
1
+ import threading
2
+ from pathlib import Path
3
+
4
+ from pilot.job.impl.base_job import BaseJob
5
+
6
+ class TabReplaceJob(BaseJob):
7
+ _begin_file_lock = threading.Lock()
8
+ def run(self):
9
+ with self._begin_file_lock:
10
+ if not self.change_current_trg_to_begin():
11
+ return
12
+ self.replace_tabs_with_spaces()
13
+ super().run()
14
+
15
+ def replace_tabs_with_spaces(self, tab_width: int = 4):
16
+ replaced_text =[]
17
+ src_path = Path(self.file_path)
18
+ spaces = ' ' * tab_width
19
+ with open(self.file_path, 'r', encoding='utf-8', newline='') as rf:
20
+ for line in rf:
21
+ replaced_text.append(line.replace('\t', spaces))
22
+
23
+ tmp_path = src_path.parent / (src_path.name + '.tmp')
24
+ with open(tmp_path, 'w', encoding='utf-8', newline='') as wf:
25
+ wf.writelines(replaced_text)
26
+
27
+ tmp_path.replace(src_path)
File without changes
@@ -0,0 +1,40 @@
1
+ import json
2
+ import time
3
+
4
+ from pilot.job.impl.base_job import BaseJob
5
+
6
+ from pilot.generater.vertexai import VertexAISingleton
7
+
8
+ class generateJsonBaseJob(BaseJob):
9
+
10
+ prompt_content: str
11
+ result_content: str
12
+ result_file_path: str
13
+
14
+ def run(self):
15
+ #with self._begin_file_lock:
16
+ # if not self.change_current_trg_to_begin():
17
+ # return
18
+ #prompt = self.get_file_content()
19
+ prompt = self.prompt_content
20
+ # トークン数チェック
21
+ vertexai = VertexAISingleton.get_instance()
22
+ token_count = vertexai.count_tokens(prompt)
23
+ if token_count == 0:
24
+ super().run()
25
+ return
26
+ if token_count > 900000:
27
+ print(f"警告: promptのトークン数が900000を超えています ({token_count} tokens)")
28
+ super().run()
29
+ return
30
+ # VertexAI で生成
31
+ start = time.time()
32
+ result = vertexai.generate_content(prompt)
33
+ end = time.time()
34
+ print(f"Ai 処理時間 {self.file_path}: {end - start:.2f}秒")
35
+
36
+ result_content = result.get('response', '')
37
+ data = json.loads(result_content)
38
+ with open(self.result_file_path, 'w', encoding='utf-8') as f:
39
+ json.dump(data, f, ensure_ascii=False, indent=2)
40
+ super().run()
@@ -0,0 +1,37 @@
1
+ import time
2
+
3
+ from pilot.job.impl.base_job import BaseJob
4
+
5
+ from pilot.generater.vertexai import VertexAISingleton
6
+
7
+ class generateTextBaseJob(BaseJob):
8
+
9
+ prompt_content: str
10
+ result_content: str
11
+ result_file_path: str
12
+
13
+ def run(self):
14
+ #with self._begin_file_lock:
15
+ # if not self.change_current_trg_to_begin():
16
+ # return
17
+ #prompt = self.get_file_content()
18
+ prompt = self.prompt_content
19
+ # トークン数チェック
20
+ vertexai = VertexAISingleton.get_instance()
21
+ token_count = vertexai.count_tokens(prompt)
22
+ if token_count == 0:
23
+ super().run()
24
+ return
25
+ if token_count > 900000:
26
+ print(f"警告: promptのトークン数が900000を超えています ({token_count} tokens)")
27
+ super().run()
28
+ return
29
+ # VertexAI で生成
30
+ start = time.time()
31
+ result = vertexai.generate_content(prompt)
32
+ end = time.time()
33
+ print(f"AI 処理時間 {self.file_path}: {end - start:.2f}秒")
34
+ result_content = result.get('response', '')
35
+ with open(self.result_file_path, 'w', encoding='utf-8') as f:
36
+ f.write(result_content)
37
+ super().run()
File without changes
File without changes
File without changes
@@ -66,9 +66,6 @@ class BaseJob(JobInterface):
66
66
  def current_trg_file_path(self, value):
67
67
  self._trg_file_path = value
68
68
 
69
-
70
-
71
-
72
69
  def run(self):
73
70
  pass
74
71
 
@@ -158,7 +155,7 @@ class BaseJob(JobInterface):
158
155
  return True
159
156
  except Exception:
160
157
  # 例外が発生した場合は False を返す
161
- print("!!!!!!!!!!!!!!!!!!!!change_current_trg_to_end erro")
158
+ #print("!!!!!!!!!!!!!!!!!!!!change_current_trg_to_end erro")
162
159
  return False
163
160
  # trgファイルが存在しなければ何もしないので、そのままreturn
164
161
  return False
@@ -341,7 +338,7 @@ class BaseJob(JobInterface):
341
338
  ジョブ実行後の後処理を行うメソッド。
342
339
  必要に応じてサブクラスでオーバーライドして使用する。
343
340
  """
344
- pass
341
+ self.change_current_trg_to_end()
345
342
 
346
343
  def generate_basedir_file(self, ext):
347
344
  dir_path = os.path.dirname(self.file_path)
@@ -31,6 +31,7 @@ class BaseUnit(UnitInterface):
31
31
  job.file_path = file_path
32
32
  if self.job_need_run(job, filename, index):
33
33
  job.run()
34
+ job.post_run()
34
35
 
35
36
  def job_need_run(self, job:BaseJob,filename: str,index):
36
37
  return True
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: pilot.linkstec
3
- Version: 0.0.90
3
+ Version: 0.0.92
4
4
  Summary: pilot of the ship, a tool for managing and deploying Python projects.
5
5
  Author-email: wanglr <wanglr1980@gmail.com>
6
6
  License-Expression: MIT
@@ -9,8 +9,10 @@ pilot/base/file_operation.py,sha256=SEguSMtkwKnOLX1KgMmFSLlPjm2IOhgUuRy07fKFCxk,
9
9
  pilot/base/get_file_encoding.py,sha256=U118_SxBXz69hhY8FGZWNQDxB1zfMG5_7gkAb6O3uG0,400
10
10
  pilot/base/make_parsing_java_file_order_base.py,sha256=mK4tIN2EIHPm7CkPrYkfVI2xA8nW1kKPsGWC4A9zat4,6310
11
11
  pilot/base/split_file_base.py,sha256=CuPI9XHlSTTkLRNXW6SNEKkdSZBOMuy685e2TcnWPkY,12625
12
+ pilot/client/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
13
+ pilot/client/ai_client.py,sha256=0n_YbjEb_QbgDAnUt_d1fJ4xYVzZmMxfJBNE5Yff9io,2919
12
14
  pilot/config/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
13
- pilot/config/config_reader.py,sha256=MchG-9j-Xfy8AquL_6tlvqQR65MFJ3c3hwWBl6IF2jw,3750
15
+ pilot/config/config_reader.py,sha256=DxTkCft9s9iuVoApWJ276ATIgM5vSiYjha4nhRVUuM0,4703
14
16
  pilot/control/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
15
17
  pilot/control/control_interface.py,sha256=zGv380oQgAKPAIHDHeFdPYzhj2Ngo2T66NWlNloA7vY,124
16
18
  pilot/control/impl/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -40,11 +42,21 @@ pilot/file_tool/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
40
42
  pilot/file_tool/create_prompt_file.py,sha256=pvmBm_iRWRvm5Qr8gaZwSAzJ_1vrqG9Xi_UJ-NgdwV8,1994
41
43
  pilot/file_tool/json_file_tool.py,sha256=v-qVcyKVhFfwEg4uJWDFMZU0EWvOwkYhrwhm-obMlUU,3637
42
44
  pilot/generater/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
43
- pilot/generater/vertexai.py,sha256=SOFfl0qCuLhfjeIoSU7Tk-I7ZB6ZrOyGme2rXDYCGzk,2599
45
+ pilot/generater/vertexai.py,sha256=gkf8ToDVGfD2j7JbR1ZtVurM7D7aZ0HLP6Vx7-Vjy7c,3939
44
46
  pilot/job/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
45
47
  pilot/job/job_interface.py,sha256=EKtuj0IcdolP494aAgTtctgamyQIoFXVwRORwOQck7A,124
48
+ pilot/job/base/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
49
+ pilot/job/base/convert/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
50
+ pilot/job/base/convert/encodingTransformerJob.py,sha256=tJI2JcBK3869pIgJxf9WFa7D-wCnHHO1aU8vQo7qjds,457
51
+ pilot/job/base/convert/tabReplaceJob.py,sha256=ev5uoPbiwhVCSDPdonvFDnWI-vi4R-4Adoa4eJddL9w,917
52
+ pilot/job/base/generater/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
53
+ pilot/job/base/generater/generateJsonBaseJob.py,sha256=Nea0GUvILgYyBiFpMOb3ZctYlxbE0pwGeJfeEWFjveQ,1348
54
+ pilot/job/base/generater/generateTextBaseJob.py,sha256=UCEkoDGN9qdHPSXrP_zpFDOpHtfjse1SskMuS31uOwk,1265
55
+ pilot/job/base/program/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
56
+ pilot/job/base/program/cobol/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
57
+ pilot/job/base/program/cobol/convert/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
46
58
  pilot/job/impl/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
47
- pilot/job/impl/base_job.py,sha256=7jxh4wREvlhFruCHA0VTnBrcg1f1sleMYo5gcvK8hdg,14642
59
+ pilot/job/impl/base_job.py,sha256=wTrAp9P6AxyQ98cD6kWa22gUTiXJc6TcTv7-cvize9c,14665
48
60
  pilot/logging/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
49
61
  pilot/logging/logger.py,sha256=TF7eGr3w8GK5v4sf71lDt97uVoBtCgqrZuCdbMmeQBU,1815
50
62
  pilot/processor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -56,11 +68,11 @@ pilot/splitters/cobolsplitter.py,sha256=oPwxKRjA7TyXWaWV3jdy59lJZy1mRn6yxD9ivqFY
56
68
  pilot/unit/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
57
69
  pilot/unit/unit_interface.py,sha256=fE8N4h_rZU-dWLHy9o0EE3yyErGmRyIuGUDb-zqe7qo,167
58
70
  pilot/unit/impl/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
59
- pilot/unit/impl/base_unit.py,sha256=LsFPpL28aSNv5rsZhfKv6CWhAw1XR4n-A6FOn2RBrZo,1272
71
+ pilot/unit/impl/base_unit.py,sha256=pF3R1TxoJtDIt5MCggSPO1H5jFI-wO1gQosz4haYTQo,1308
60
72
  pilot/util/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
61
73
  pilot/util/files.py,sha256=v9uzfzo3Aq4xgnUIASEZeBJoA2nD9Qz_EA3P-FwzGFQ,1896
62
- pilot_linkstec-0.0.90.dist-info/licenses/LICENSE,sha256=6kbiFSfobTZ7beWiKnHpN902HgBx-Jzgcme0SvKqhKY,1091
63
- pilot_linkstec-0.0.90.dist-info/METADATA,sha256=iSPnwKqO0yHZW8_WqwTvUl_Ta6HUmJkXZcbJjOxMSS4,679
64
- pilot_linkstec-0.0.90.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
65
- pilot_linkstec-0.0.90.dist-info/top_level.txt,sha256=BijnVJdXnIPxxx3s60M848seL4Z12gNUPod6KPJxK9c,6
66
- pilot_linkstec-0.0.90.dist-info/RECORD,,
74
+ pilot_linkstec-0.0.92.dist-info/licenses/LICENSE,sha256=6kbiFSfobTZ7beWiKnHpN902HgBx-Jzgcme0SvKqhKY,1091
75
+ pilot_linkstec-0.0.92.dist-info/METADATA,sha256=wybOFFLfloJoPHK_ZTdcmDl5-PiR-esNp_N1GajX_tM,679
76
+ pilot_linkstec-0.0.92.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
77
+ pilot_linkstec-0.0.92.dist-info/top_level.txt,sha256=BijnVJdXnIPxxx3s60M848seL4Z12gNUPod6KPJxK9c,6
78
+ pilot_linkstec-0.0.92.dist-info/RECORD,,