isa-model 0.0.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (86) hide show
  1. isa_model/__init__.py +5 -0
  2. isa_model/core/model_manager.py +143 -0
  3. isa_model/core/model_registry.py +115 -0
  4. isa_model/core/model_router.py +226 -0
  5. isa_model/core/model_storage.py +133 -0
  6. isa_model/core/model_version.py +0 -0
  7. isa_model/core/resource_manager.py +202 -0
  8. isa_model/core/storage/hf_storage.py +0 -0
  9. isa_model/core/storage/local_storage.py +0 -0
  10. isa_model/core/storage/minio_storage.py +0 -0
  11. isa_model/deployment/gpu_fp16_ds8/models/deepseek_r1/1/model.py +120 -0
  12. isa_model/deployment/gpu_fp16_ds8/scripts/download_model.py +18 -0
  13. isa_model/deployment/gpu_int8_ds8/app/server.py +66 -0
  14. isa_model/deployment/gpu_int8_ds8/scripts/test_client.py +43 -0
  15. isa_model/deployment/gpu_int8_ds8/scripts/test_client_os.py +35 -0
  16. isa_model/inference/__init__.py +11 -0
  17. isa_model/inference/adapter/unified_api.py +248 -0
  18. isa_model/inference/ai_factory.py +359 -0
  19. isa_model/inference/base.py +46 -0
  20. isa_model/inference/providers/__init__.py +19 -0
  21. isa_model/inference/providers/base_provider.py +30 -0
  22. isa_model/inference/providers/model_cache_manager.py +341 -0
  23. isa_model/inference/providers/ollama_provider.py +73 -0
  24. isa_model/inference/providers/openai_provider.py +101 -0
  25. isa_model/inference/providers/replicate_provider.py +107 -0
  26. isa_model/inference/providers/triton_provider.py +439 -0
  27. isa_model/inference/services/__init__.py +14 -0
  28. isa_model/inference/services/audio/base_stt_service.py +91 -0
  29. isa_model/inference/services/audio/base_tts_service.py +136 -0
  30. isa_model/inference/services/audio/openai_tts_service.py +71 -0
  31. isa_model/inference/services/base_service.py +106 -0
  32. isa_model/inference/services/embedding/ollama_embed_service.py +97 -0
  33. isa_model/inference/services/embedding/openai_embed_service.py +0 -0
  34. isa_model/inference/services/llm/__init__.py +12 -0
  35. isa_model/inference/services/llm/base_llm_service.py +134 -0
  36. isa_model/inference/services/llm/ollama_llm_service.py +99 -0
  37. isa_model/inference/services/llm/openai_llm_service.py +138 -0
  38. isa_model/inference/services/others/table_transformer_service.py +61 -0
  39. isa_model/inference/services/vision/__init__.py +12 -0
  40. isa_model/inference/services/vision/helpers/image_utils.py +58 -0
  41. isa_model/inference/services/vision/helpers/text_splitter.py +46 -0
  42. isa_model/inference/services/vision/ollama_vision_service.py +60 -0
  43. isa_model/inference/services/vision/openai_vision_service.py +80 -0
  44. isa_model/inference/services/vision/replicate_image_gen_service.py +185 -0
  45. isa_model/inference/utils/conversion/bge_rerank_convert.py +73 -0
  46. isa_model/inference/utils/conversion/onnx_converter.py +0 -0
  47. isa_model/inference/utils/conversion/torch_converter.py +0 -0
  48. isa_model/scripts/inference_tracker.py +283 -0
  49. isa_model/scripts/mlflow_manager.py +379 -0
  50. isa_model/scripts/model_registry.py +465 -0
  51. isa_model/scripts/start_mlflow.py +95 -0
  52. isa_model/scripts/training_tracker.py +257 -0
  53. isa_model/training/engine/llama_factory/__init__.py +39 -0
  54. isa_model/training/engine/llama_factory/config.py +115 -0
  55. isa_model/training/engine/llama_factory/data_adapter.py +284 -0
  56. isa_model/training/engine/llama_factory/examples/__init__.py +6 -0
  57. isa_model/training/engine/llama_factory/examples/finetune_with_tracking.py +185 -0
  58. isa_model/training/engine/llama_factory/examples/rlhf_with_tracking.py +163 -0
  59. isa_model/training/engine/llama_factory/factory.py +331 -0
  60. isa_model/training/engine/llama_factory/rl.py +254 -0
  61. isa_model/training/engine/llama_factory/trainer.py +171 -0
  62. isa_model/training/image_model/configs/create_config.py +37 -0
  63. isa_model/training/image_model/configs/create_flux_config.py +26 -0
  64. isa_model/training/image_model/configs/create_lora_config.py +21 -0
  65. isa_model/training/image_model/prepare_massed_compute.py +97 -0
  66. isa_model/training/image_model/prepare_upload.py +17 -0
  67. isa_model/training/image_model/raw_data/create_captions.py +16 -0
  68. isa_model/training/image_model/raw_data/create_lora_captions.py +20 -0
  69. isa_model/training/image_model/raw_data/pre_processing.py +200 -0
  70. isa_model/training/image_model/train/train.py +42 -0
  71. isa_model/training/image_model/train/train_flux.py +41 -0
  72. isa_model/training/image_model/train/train_lora.py +57 -0
  73. isa_model/training/image_model/train_main.py +25 -0
  74. isa_model/training/llm_model/annotation/annotation_schema.py +47 -0
  75. isa_model/training/llm_model/annotation/processors/annotation_processor.py +126 -0
  76. isa_model/training/llm_model/annotation/storage/dataset_manager.py +131 -0
  77. isa_model/training/llm_model/annotation/storage/dataset_schema.py +44 -0
  78. isa_model/training/llm_model/annotation/tests/test_annotation_flow.py +109 -0
  79. isa_model/training/llm_model/annotation/tests/test_minio copy.py +113 -0
  80. isa_model/training/llm_model/annotation/tests/test_minio_upload.py +43 -0
  81. isa_model/training/llm_model/annotation/views/annotation_controller.py +158 -0
  82. isa_model-0.0.1.dist-info/METADATA +327 -0
  83. isa_model-0.0.1.dist-info/RECORD +86 -0
  84. isa_model-0.0.1.dist-info/WHEEL +5 -0
  85. isa_model-0.0.1.dist-info/licenses/LICENSE +21 -0
  86. isa_model-0.0.1.dist-info/top_level.txt +1 -0
@@ -0,0 +1,44 @@
1
+ # app/services/llm_model/annotation/dataset/dataset_schema.py
2
+ from enum import Enum
3
+ from pydantic import BaseModel, Field
4
+ from typing import Dict, List, Optional
5
+ from datetime import datetime
6
+ from bson import ObjectId
7
+
8
+ class DatasetType(str, Enum):
9
+ SFT = "sft"
10
+ RLHF = "rlhf"
11
+
12
+ class DatasetStatus(str, Enum):
13
+ PENDING = "pending"
14
+ PROCESSING = "processing"
15
+ READY = "ready"
16
+ ERROR = "error"
17
+
18
+ class DatasetFiles(BaseModel):
19
+ train: str
20
+ eval: Optional[str]
21
+ test: Optional[str]
22
+
23
+ class DatasetStats(BaseModel):
24
+ total_examples: int
25
+ avg_length: Optional[float]
26
+ num_conversations: Optional[int]
27
+ additional_metrics: Optional[Dict] = {}
28
+
29
+ class Dataset(BaseModel):
30
+ id: Optional[ObjectId] = Field(None, alias="_id")
31
+ name: str
32
+ type: DatasetType
33
+ version: str
34
+ storage_path: str
35
+ files: DatasetFiles
36
+ stats: DatasetStats
37
+ source_annotations: List[str]
38
+ created_at: datetime
39
+ status: DatasetStatus
40
+ metadata: Optional[Dict] = {}
41
+
42
+ class Config:
43
+ arbitrary_types_allowed = True
44
+ populate_by_name = True
@@ -0,0 +1,109 @@
1
+ # test_annotation_flow.py
2
+ import os
3
+ os.environ["ENV"] = "local"
4
+
5
+ import asyncio
6
+ from datetime import datetime
7
+ from bson import ObjectId
8
+ from app.services.llm_model.annotation.views.annotation_controller import AnnotationController
9
+ from app.services.llm_model.annotation.processors.annotation_processor import AnnotationProcessor
10
+ from app.services.llm_model.annotation.annotation_schema import (
11
+ AnnotationFeedback,
12
+ RatingScale,
13
+ AnnotationType,
14
+ AnnotationAspects,
15
+ BetterResponse
16
+ )
17
+ from app.config.config_manager import config_manager
18
+
19
+ async def setup_test_data():
20
+ """Setup initial test data in MongoDB"""
21
+ db = await config_manager.get_db('mongodb')
22
+
23
+ # Create a test annotation
24
+ test_annotation = {
25
+ "_id": ObjectId(),
26
+ "project_name": "test_project",
27
+ "items": [{
28
+ "item_id": "test_item_1",
29
+ "input": {
30
+ "messages": [{
31
+ "role": "user",
32
+ "content": "What is the capital of France?"
33
+ }]
34
+ },
35
+ "output": {
36
+ "content": "The capital of France is Paris."
37
+ },
38
+ "status": "pending"
39
+ }],
40
+ "created_at": datetime.utcnow().isoformat()
41
+ }
42
+
43
+ await db['annotations'].insert_one(test_annotation)
44
+ return test_annotation
45
+
46
+ async def test_annotation_flow():
47
+ """Test the complete annotation flow"""
48
+ try:
49
+ # Initialize controllers
50
+ annotation_controller = AnnotationController()
51
+ annotation_processor = AnnotationProcessor()
52
+
53
+ # Setup test data
54
+ test_data = await setup_test_data()
55
+ annotation_id = str(test_data["_id"])
56
+ item_id = test_data["items"][0]["item_id"]
57
+
58
+ print("1. Created test annotation")
59
+
60
+ # Create test feedback
61
+ feedback = AnnotationFeedback(
62
+ rating=RatingScale.EXCELLENT,
63
+ category=AnnotationType.ACCURACY,
64
+ aspects=AnnotationAspects(
65
+ factually_correct=True,
66
+ relevant=True,
67
+ harmful=False,
68
+ biased=False,
69
+ complete=True,
70
+ efficient=True
71
+ ),
72
+ better_response=BetterResponse(
73
+ content="Paris is the capital city of France, known for its iconic Eiffel Tower.",
74
+ reason="Added more context and detail"
75
+ ),
76
+ comment="Good response, but could be more detailed"
77
+ )
78
+
79
+ # Submit annotation
80
+ result = await annotation_controller.submit_annotation(
81
+ annotation_id=annotation_id,
82
+ item_id=item_id,
83
+ feedback=feedback,
84
+ annotator_id="test_annotator"
85
+ )
86
+
87
+ print("2. Submitted annotation:", result)
88
+
89
+ # Process annotation queue
90
+ await annotation_processor.process_queue()
91
+ print("3. Processed annotation queue")
92
+
93
+ # Verify dataset creation
94
+ db = await config_manager.get_db('mongodb')
95
+ datasets = await db['training_datasets'].find().to_list(length=10)
96
+
97
+ print("\nCreated Datasets:")
98
+ for dataset in datasets:
99
+ print(f"- {dataset['name']} ({dataset['type']})")
100
+ print(f" Status: {dataset['status']}")
101
+ print(f" Examples: {dataset['stats']['total_examples']}")
102
+
103
+ except Exception as e:
104
+ print(f"Error during test: {e}")
105
+
106
+ if __name__ == "__main__":
107
+ # Run the test
108
+ print("Starting annotation flow test...")
109
+ asyncio.run(test_annotation_flow())
@@ -0,0 +1,113 @@
1
+ #!/usr/bin/env python3
2
+
3
+ import os
4
+ from minio import Minio
5
+ import json
6
+ import logging
7
+ from io import BytesIO
8
+
9
+ # Configure logging
10
+ logging.basicConfig(level=logging.INFO)
11
+ logger = logging.getLogger(__name__)
12
+
13
+ def test_minio():
14
+ try:
15
+ # Get MinIO host from environment variable or use default
16
+ minio_host = os.getenv("MINIO_HOST", "localhost:9000")
17
+ logger.info(f"Using MinIO host: {minio_host}")
18
+
19
+ # 1. Create MinIO client
20
+ logger.info("Creating MinIO client...")
21
+ client = Minio(
22
+ minio_host,
23
+ access_key="minioadmin",
24
+ secret_key="minioadmin",
25
+ secure=False
26
+ )
27
+
28
+ # 2. Test bucket operations
29
+ bucket_name = "knowledge-files" # Changed to match the actual bucket name
30
+ logger.info(f"Testing bucket operations with {bucket_name}...")
31
+
32
+ # Create bucket if it doesn't exist
33
+ if not client.bucket_exists(bucket_name):
34
+ client.make_bucket(bucket_name)
35
+ logger.info(f"Created new bucket: {bucket_name}")
36
+ else:
37
+ logger.info(f"Using existing bucket: {bucket_name}")
38
+
39
+ # Set bucket policy - allow all operations
40
+ policy = {
41
+ "Version": "2012-10-17",
42
+ "Statement": [
43
+ {
44
+ "Effect": "Allow",
45
+ "Principal": {"AWS": ["*"]},
46
+ "Action": [
47
+ "s3:GetBucketLocation",
48
+ "s3:ListBucket",
49
+ "s3:ListBucketMultipartUploads"
50
+ ],
51
+ "Resource": [f"arn:aws:s3:::{bucket_name}"]
52
+ },
53
+ {
54
+ "Effect": "Allow",
55
+ "Principal": {"AWS": ["*"]},
56
+ "Action": [
57
+ "s3:AbortMultipartUpload",
58
+ "s3:DeleteObject",
59
+ "s3:GetObject",
60
+ "s3:ListMultipartUploadParts",
61
+ "s3:PutObject"
62
+ ],
63
+ "Resource": [f"arn:aws:s3:::{bucket_name}/*"]
64
+ }
65
+ ]
66
+ }
67
+
68
+ try:
69
+ client.set_bucket_policy(bucket_name, json.dumps(policy))
70
+ logger.info("Set bucket policy successfully")
71
+ except Exception as e:
72
+ logger.warning(f"Failed to set bucket policy: {e}")
73
+
74
+ # 3. Test file upload
75
+ source_file = "init-scripts/files/haley_system.txt"
76
+ object_name = "haley_system.txt"
77
+
78
+ if os.path.exists(source_file):
79
+ # Get file size
80
+ file_size = os.path.getsize(source_file)
81
+ logger.info(f"Found source file: {source_file} (size: {file_size} bytes)")
82
+
83
+ # Upload file
84
+ with open(source_file, 'rb') as file_data:
85
+ client.put_object(
86
+ bucket_name,
87
+ object_name,
88
+ file_data,
89
+ file_size,
90
+ content_type="text/plain"
91
+ )
92
+ logger.info(f"Uploaded file: {object_name}")
93
+
94
+ # 4. Test file download
95
+ data = client.get_object(bucket_name, object_name)
96
+ content = data.read().decode('utf-8')
97
+ logger.info(f"Successfully downloaded file. First 100 chars: {content[:100]}...")
98
+
99
+ # 5. Verify file exists
100
+ stat = client.stat_object(bucket_name, object_name)
101
+ logger.info(f"File stats: {stat}")
102
+
103
+ logger.info("Test completed successfully!")
104
+ else:
105
+ logger.error(f"Source file not found: {source_file}")
106
+ raise FileNotFoundError(f"Source file not found: {source_file}")
107
+
108
+ except Exception as e:
109
+ logger.error(f"Test failed: {str(e)}")
110
+ raise
111
+
112
+ if __name__ == "__main__":
113
+ test_minio()
@@ -0,0 +1,43 @@
1
+ #!/usr/bin/env python3
2
+ import os
3
+ from minio import Minio
4
+ import io
5
+
6
+ # MinIO client setup
7
+ client = Minio(
8
+ "localhost:9000",
9
+ access_key="minioadmin",
10
+ secret_key="minioadmin",
11
+ secure=False
12
+ )
13
+
14
+ # Test bucket name
15
+ bucket_name = "test-bucket"
16
+
17
+ # Create bucket if not exists
18
+ if not client.bucket_exists(bucket_name):
19
+ client.make_bucket(bucket_name)
20
+ print(f"Created bucket: {bucket_name}")
21
+ else:
22
+ print(f"Bucket already exists: {bucket_name}")
23
+
24
+ # Upload test file
25
+ source_file = "init-scripts/files/haley_system.txt"
26
+ if os.path.exists(source_file):
27
+ with open(source_file, 'rb') as file_data:
28
+ content = file_data.read()
29
+ # Upload file
30
+ client.put_object(
31
+ bucket_name,
32
+ "haley_system.txt",
33
+ io.BytesIO(content),
34
+ len(content),
35
+ content_type="text/plain"
36
+ )
37
+ print(f"Successfully uploaded {source_file}")
38
+
39
+ # Verify upload
40
+ stat = client.stat_object(bucket_name, "haley_system.txt")
41
+ print(f"File stats: {stat}")
42
+ else:
43
+ print(f"Source file not found: {source_file}")
@@ -0,0 +1,158 @@
1
+ # app/services/llm_model/tracing/annotation/annotation_controller.py
2
+ from typing import Dict, Any, List, Optional
3
+ from datetime import datetime
4
+ from bson import ObjectId
5
+ from app.config.config_manager import config_manager
6
+ from app.services.training.llm_model.annotation.annotation_schema import AnnotationFeedback, RatingScale
7
+ from app.services.training.llm_model.annotation.storage.dataset_manager import DatasetManager
8
+
9
+
10
+ class AnnotationController:
11
+ def __init__(self):
12
+ self.logger = config_manager.get_logger(__name__)
13
+
14
+ async def get_pending_annotations(
15
+ self,
16
+ project_name: str,
17
+ category: Optional[str] = None,
18
+ min_rating: Optional[int] = None,
19
+ page: int = 1,
20
+ limit: int = 10
21
+ ) -> Dict[str, Any]:
22
+ """Get filtered list of pending annotations"""
23
+ db = await config_manager.get_db('mongodb')
24
+ collection = db['annotations']
25
+
26
+ # Build query with filters
27
+ query = {"status": "pending", "project_name": project_name}
28
+ if category:
29
+ query["annotation_type"] = category
30
+ if min_rating:
31
+ query["items.feedback.rating"] = {"$gte": min_rating}
32
+
33
+ annotations = await collection.find(query)\
34
+ .sort("created_at", -1)\
35
+ .skip((page - 1) * limit)\
36
+ .limit(limit)\
37
+ .to_list(length=limit)
38
+
39
+ return {
40
+ "annotations": annotations,
41
+ "pagination": {
42
+ "page": page,
43
+ "limit": limit,
44
+ "total": await collection.count_documents(query)
45
+ }
46
+ }
47
+
48
+ async def submit_annotation(
49
+ self,
50
+ annotation_id: str,
51
+ item_id: str,
52
+ feedback: AnnotationFeedback,
53
+ annotator_id: str
54
+ ) -> Dict[str, Any]:
55
+ """Submit and process annotation feedback"""
56
+ db = await config_manager.get_db('mongodb')
57
+ collection = db['annotations']
58
+
59
+ # Determine if annotation should be selected for training
60
+ is_selected = self._evaluate_for_training(feedback)
61
+ feedback_dict = feedback.dict()
62
+ feedback_dict["is_selected_for_training"] = is_selected
63
+
64
+ # Update annotation
65
+ result = await collection.update_one(
66
+ {
67
+ "_id": ObjectId(annotation_id),
68
+ "items.item_id": item_id
69
+ },
70
+ {
71
+ "$set": {
72
+ "items.$.feedback": feedback_dict,
73
+ "items.$.status": "completed",
74
+ "items.$.annotated_at": datetime.utcnow().isoformat(),
75
+ "items.$.annotator_id": annotator_id,
76
+ "items.$.training_status": "pending" if is_selected else "none"
77
+ }
78
+ }
79
+ )
80
+
81
+ # Process for training if selected
82
+ if is_selected:
83
+ await self._queue_for_training(annotation_id, item_id, feedback)
84
+
85
+ return {
86
+ "status": "success",
87
+ "selected_for_training": is_selected,
88
+ "message": "Annotation submitted successfully"
89
+ }
90
+
91
+ def _evaluate_for_training(self, feedback: AnnotationFeedback) -> bool:
92
+ """Evaluate if annotation should be used for training"""
93
+ # Select for SFT if rating is excellent and aspects are positive
94
+ if feedback.rating == RatingScale.EXCELLENT:
95
+ aspects = feedback.aspects
96
+ if all([
97
+ aspects.factually_correct,
98
+ aspects.relevant,
99
+ not aspects.harmful,
100
+ not aspects.biased
101
+ ]):
102
+ return True
103
+
104
+ # Select for RLHF if better response is provided
105
+ if feedback.better_response:
106
+ return True
107
+
108
+ return False
109
+
110
+ async def _queue_for_training(
111
+ self,
112
+ annotation_id: str,
113
+ item_id: str,
114
+ feedback: AnnotationFeedback
115
+ ):
116
+ """Queue selected annotations for training data generation"""
117
+ db = await config_manager.get_db('mongodb')
118
+ training_queue = db['training_queue']
119
+
120
+ await training_queue.insert_one({
121
+ "annotation_id": annotation_id,
122
+ "item_id": item_id,
123
+ "type": "sft" if feedback.rating == RatingScale.EXCELLENT else "rlhf",
124
+ "feedback": feedback.dict(),
125
+ "status": "pending",
126
+ "created_at": datetime.utcnow().isoformat()
127
+ })
128
+
129
+ class DatasetPreparationProcessor:
130
+ def __init__(self):
131
+ self.logger = config_manager.get_logger(__name__)
132
+ self.dataset_manager = DatasetManager()
133
+ self.batch_size = 1000 # Configure as needed
134
+
135
+ async def process_annotation_queue(self) -> None:
136
+ """Process pending annotations and prepare datasets"""
137
+ db = await config_manager.get_db('mongodb')
138
+ annotation_queue = db['dataset_preparation_queue']
139
+
140
+ # Process items for SFT dataset
141
+ sft_items = await self._get_pending_annotations("sft")
142
+ if len(sft_items) >= self.batch_size:
143
+ await self._create_sft_dataset(sft_items)
144
+
145
+ # Process items for RLHF dataset
146
+ rlhf_items = await self._get_pending_annotations("rlhf")
147
+ if len(rlhf_items) >= self.batch_size:
148
+ await self._create_rlhf_dataset(rlhf_items)
149
+
150
+ async def _get_pending_annotations(self, dataset_type: str) -> List[Dict[str, Any]]:
151
+ """Get pending annotations for dataset preparation"""
152
+ db = await config_manager.get_db('mongodb')
153
+ queue = db['dataset_preparation_queue']
154
+
155
+ return await queue.find({
156
+ "status": "pending",
157
+ "dataset_type": dataset_type
158
+ }).to_list(length=self.batch_size)