unstructured-ingest 0.5.8__py3-none-any.whl → 0.5.9__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of unstructured-ingest might be problematic. Click here for more details.
- test/integration/connectors/test_astradb.py +32 -5
- unstructured_ingest/__version__.py +1 -1
- unstructured_ingest/v2/processes/connectors/astradb.py +85 -13
- unstructured_ingest/v2/processes/connectors/confluence.py +2 -2
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/METADATA +23 -23
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/RECORD +10 -10
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/LICENSE.md +0 -0
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/WHEEL +0 -0
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/entry_points.txt +0 -0
- {unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/top_level.txt +0 -0
|
@@ -56,12 +56,7 @@ def test_precheck_succeeds_indexer(connection_config: AstraDBConnectionConfig):
|
|
|
56
56
|
connection_config=connection_config,
|
|
57
57
|
index_config=AstraDBIndexerConfig(collection_name=EXISTENT_COLLECTION_NAME),
|
|
58
58
|
)
|
|
59
|
-
uploader = AstraDBUploader(
|
|
60
|
-
connection_config=connection_config,
|
|
61
|
-
upload_config=AstraDBUploaderConfig(collection_name=EXISTENT_COLLECTION_NAME),
|
|
62
|
-
)
|
|
63
59
|
indexer.precheck()
|
|
64
|
-
uploader.precheck()
|
|
65
60
|
|
|
66
61
|
|
|
67
62
|
@pytest.mark.tags(CONNECTOR_TYPE, DESTINATION_TAG, VECTOR_DB_TAG)
|
|
@@ -73,6 +68,12 @@ def test_precheck_succeeds_uploader(connection_config: AstraDBConnectionConfig):
|
|
|
73
68
|
)
|
|
74
69
|
uploader.precheck()
|
|
75
70
|
|
|
71
|
+
uploader2 = AstraDBUploader(
|
|
72
|
+
connection_config=connection_config,
|
|
73
|
+
upload_config=AstraDBUploaderConfig(),
|
|
74
|
+
)
|
|
75
|
+
uploader2.precheck()
|
|
76
|
+
|
|
76
77
|
|
|
77
78
|
@pytest.mark.tags(CONNECTOR_TYPE, SOURCE_TAG, VECTOR_DB_TAG)
|
|
78
79
|
@requires_env("ASTRA_DB_APPLICATION_TOKEN", "ASTRA_DB_API_ENDPOINT")
|
|
@@ -216,6 +217,32 @@ async def test_astra_search_destination(
|
|
|
216
217
|
)
|
|
217
218
|
|
|
218
219
|
|
|
220
|
+
@pytest.mark.tags(CONNECTOR_TYPE, DESTINATION_TAG, VECTOR_DB_TAG)
|
|
221
|
+
@requires_env("ASTRA_DB_API_ENDPOINT", "ASTRA_DB_APPLICATION_TOKEN")
|
|
222
|
+
def test_astra_create_destination():
|
|
223
|
+
env_data = get_env_data()
|
|
224
|
+
connection_config = AstraDBConnectionConfig(
|
|
225
|
+
access_config=AstraDBAccessConfig(api_endpoint=env_data.api_endpoint, token=env_data.token),
|
|
226
|
+
)
|
|
227
|
+
uploader = AstraDBUploader(
|
|
228
|
+
connection_config=connection_config,
|
|
229
|
+
upload_config=AstraDBUploaderConfig(),
|
|
230
|
+
)
|
|
231
|
+
collection_name = "system_created-123"
|
|
232
|
+
formatted_collection_name = "system_created_123"
|
|
233
|
+
created = uploader.create_destination(destination_name=collection_name, vector_length=3072)
|
|
234
|
+
assert created
|
|
235
|
+
assert uploader.upload_config.collection_name == formatted_collection_name
|
|
236
|
+
|
|
237
|
+
created = uploader.create_destination(destination_name=collection_name, vector_length=3072)
|
|
238
|
+
assert not created
|
|
239
|
+
|
|
240
|
+
# cleanup
|
|
241
|
+
client = AstraDBClient()
|
|
242
|
+
db = client.get_database(api_endpoint=env_data.api_endpoint, token=env_data.token)
|
|
243
|
+
db.drop_collection(formatted_collection_name)
|
|
244
|
+
|
|
245
|
+
|
|
219
246
|
@pytest.mark.tags(CONNECTOR_TYPE, DESTINATION_TAG, VECTOR_DB_TAG)
|
|
220
247
|
@pytest.mark.parametrize("upload_file_str", ["upload_file_ndjson", "upload_file"])
|
|
221
248
|
def test_astra_stager(
|
|
@@ -1 +1 @@
|
|
|
1
|
-
__version__ = "0.5.
|
|
1
|
+
__version__ = "0.5.9" # pragma: no cover
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import csv
|
|
2
2
|
import hashlib
|
|
3
|
+
import re
|
|
3
4
|
from dataclasses import dataclass, field
|
|
4
5
|
from pathlib import Path
|
|
5
6
|
from time import time
|
|
@@ -48,6 +49,7 @@ if TYPE_CHECKING:
|
|
|
48
49
|
from astrapy import AsyncCollection as AstraDBAsyncCollection
|
|
49
50
|
from astrapy import Collection as AstraDBCollection
|
|
50
51
|
from astrapy import DataAPIClient as AstraDBClient
|
|
52
|
+
from astrapy import Database as AstraDB
|
|
51
53
|
|
|
52
54
|
|
|
53
55
|
CONNECTOR_TYPE = "astradb"
|
|
@@ -85,11 +87,10 @@ class AstraDBConnectionConfig(ConnectionConfig):
|
|
|
85
87
|
)
|
|
86
88
|
|
|
87
89
|
|
|
88
|
-
def
|
|
90
|
+
def get_astra_db(
|
|
89
91
|
connection_config: AstraDBConnectionConfig,
|
|
90
|
-
collection_name: str,
|
|
91
92
|
keyspace: str,
|
|
92
|
-
) -> "
|
|
93
|
+
) -> "AstraDB":
|
|
93
94
|
# Build the Astra DB object.
|
|
94
95
|
access_configs = connection_config.access_config.get_secret_value()
|
|
95
96
|
|
|
@@ -103,9 +104,20 @@ def get_astra_collection(
|
|
|
103
104
|
token=access_configs.token,
|
|
104
105
|
keyspace=keyspace,
|
|
105
106
|
)
|
|
107
|
+
return astra_db
|
|
108
|
+
|
|
106
109
|
|
|
107
|
-
|
|
110
|
+
def get_astra_collection(
|
|
111
|
+
connection_config: AstraDBConnectionConfig,
|
|
112
|
+
collection_name: str,
|
|
113
|
+
keyspace: str,
|
|
114
|
+
) -> "AstraDBCollection":
|
|
115
|
+
|
|
116
|
+
astra_db = get_astra_db(connection_config=connection_config, keyspace=keyspace)
|
|
117
|
+
|
|
118
|
+
# astradb will return a collection object in all cases (even if it doesn't exist)
|
|
108
119
|
astra_db_collection = astra_db.get_collection(name=collection_name)
|
|
120
|
+
|
|
109
121
|
return astra_db_collection
|
|
110
122
|
|
|
111
123
|
|
|
@@ -151,10 +163,11 @@ class AstraDBDownloaderConfig(DownloaderConfig):
|
|
|
151
163
|
|
|
152
164
|
|
|
153
165
|
class AstraDBUploaderConfig(UploaderConfig):
|
|
154
|
-
collection_name: str = Field(
|
|
166
|
+
collection_name: Optional[str] = Field(
|
|
155
167
|
description="The name of the Astra DB collection. "
|
|
156
168
|
"Note that the collection name must only include letters, "
|
|
157
|
-
"numbers, and underscores."
|
|
169
|
+
"numbers, and underscores.",
|
|
170
|
+
default=None,
|
|
158
171
|
)
|
|
159
172
|
keyspace: Optional[str] = Field(default=None, description="The Astra DB connection keyspace.")
|
|
160
173
|
requested_indexing_policy: Optional[dict[str, Any]] = Field(
|
|
@@ -337,25 +350,84 @@ class AstraDBUploader(Uploader):
|
|
|
337
350
|
upload_config: AstraDBUploaderConfig
|
|
338
351
|
connector_type: str = CONNECTOR_TYPE
|
|
339
352
|
|
|
353
|
+
def init(self, **kwargs: Any) -> None:
|
|
354
|
+
self.create_destination(**kwargs)
|
|
355
|
+
|
|
340
356
|
def precheck(self) -> None:
|
|
341
357
|
try:
|
|
342
|
-
|
|
343
|
-
|
|
344
|
-
|
|
345
|
-
|
|
346
|
-
|
|
358
|
+
if self.upload_config.collection_name:
|
|
359
|
+
self.get_collection(collection_name=self.upload_config.collection_name).options()
|
|
360
|
+
else:
|
|
361
|
+
# check for db connection only if collection name is not provided
|
|
362
|
+
get_astra_db(
|
|
363
|
+
connection_config=self.connection_config,
|
|
364
|
+
keyspace=self.upload_config.keyspace,
|
|
365
|
+
)
|
|
347
366
|
except Exception as e:
|
|
348
367
|
logger.error(f"Failed to validate connection {e}", exc_info=True)
|
|
349
368
|
raise DestinationConnectionError(f"failed to validate connection: {e}")
|
|
350
369
|
|
|
351
370
|
@requires_dependencies(["astrapy"], extras="astradb")
|
|
352
|
-
def get_collection(self) -> "AstraDBCollection":
|
|
371
|
+
def get_collection(self, collection_name: Optional[str] = None) -> "AstraDBCollection":
|
|
353
372
|
return get_astra_collection(
|
|
354
373
|
connection_config=self.connection_config,
|
|
355
|
-
collection_name=self.upload_config.collection_name,
|
|
374
|
+
collection_name=collection_name or self.upload_config.collection_name,
|
|
356
375
|
keyspace=self.upload_config.keyspace,
|
|
357
376
|
)
|
|
358
377
|
|
|
378
|
+
def _collection_exists(self, collection_name: str):
|
|
379
|
+
from astrapy.exceptions import CollectionNotFoundException
|
|
380
|
+
|
|
381
|
+
collection = get_astra_collection(
|
|
382
|
+
connection_config=self.connection_config,
|
|
383
|
+
collection_name=collection_name,
|
|
384
|
+
keyspace=self.upload_config.keyspace,
|
|
385
|
+
)
|
|
386
|
+
|
|
387
|
+
try:
|
|
388
|
+
collection.options()
|
|
389
|
+
return True
|
|
390
|
+
except CollectionNotFoundException:
|
|
391
|
+
return False
|
|
392
|
+
except Exception as e:
|
|
393
|
+
logger.error(f"failed to check if astra collection exists : {e}")
|
|
394
|
+
raise DestinationConnectionError(f"failed to check if astra collection exists : {e}")
|
|
395
|
+
|
|
396
|
+
def format_destination_name(self, destination_name: str) -> str:
|
|
397
|
+
# AstraDB collection naming requirements:
|
|
398
|
+
# must be below 50 characters
|
|
399
|
+
# must be lowercase alphanumeric and underscores only
|
|
400
|
+
formatted = re.sub(r"[^a-z0-9]", "_", destination_name.lower())
|
|
401
|
+
return formatted
|
|
402
|
+
|
|
403
|
+
def create_destination(
|
|
404
|
+
self,
|
|
405
|
+
vector_length: int,
|
|
406
|
+
destination_name: str = "unstructuredautocreated",
|
|
407
|
+
similarity_metric: Optional[str] = "cosine",
|
|
408
|
+
**kwargs: Any,
|
|
409
|
+
) -> bool:
|
|
410
|
+
destination_name = self.format_destination_name(destination_name)
|
|
411
|
+
collection_name = self.upload_config.collection_name or destination_name
|
|
412
|
+
self.upload_config.collection_name = collection_name
|
|
413
|
+
|
|
414
|
+
if not self._collection_exists(collection_name):
|
|
415
|
+
astra_db = get_astra_db(
|
|
416
|
+
connection_config=self.connection_config, keyspace=self.upload_config.keyspace
|
|
417
|
+
)
|
|
418
|
+
logger.info(
|
|
419
|
+
f"creating default astra collection '{collection_name}' with dimension "
|
|
420
|
+
f"{vector_length} and metric {similarity_metric}"
|
|
421
|
+
)
|
|
422
|
+
astra_db.create_collection(
|
|
423
|
+
collection_name,
|
|
424
|
+
dimension=vector_length,
|
|
425
|
+
metric=similarity_metric,
|
|
426
|
+
)
|
|
427
|
+
return True
|
|
428
|
+
logger.debug(f"collection with name '{collection_name}' already exists, skipping creation")
|
|
429
|
+
return False
|
|
430
|
+
|
|
359
431
|
def delete_by_record_id(self, collection: "AstraDBCollection", file_data: FileData):
|
|
360
432
|
logger.debug(
|
|
361
433
|
f"deleting records from collection {collection.name} "
|
|
@@ -233,9 +233,9 @@ class ConfluenceDownloader(Downloader):
|
|
|
233
233
|
raise ValueError(f"Page with ID {doc_id} does not exist.")
|
|
234
234
|
|
|
235
235
|
content = page["body"]["view"]["value"]
|
|
236
|
-
# This supports v2 html parsing in unstructured
|
|
237
236
|
title = page["title"]
|
|
238
|
-
|
|
237
|
+
# Using h1 for title is supported by both v1 and v2 html parsing in unstructured
|
|
238
|
+
title_html = f"<h1>{title}</h1>"
|
|
239
239
|
content = f"<body class='Document' >{title_html}{content}</body>"
|
|
240
240
|
if self.download_config.extract_images:
|
|
241
241
|
with self.connection_config.get_client() as client:
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: unstructured-ingest
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.9
|
|
4
4
|
Summary: A library that prepares raw documents for downstream ML tasks.
|
|
5
5
|
Home-page: https://github.com/Unstructured-IO/unstructured-ingest
|
|
6
6
|
Author: Unstructured Technologies
|
|
@@ -22,13 +22,13 @@ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
|
|
|
22
22
|
Requires-Python: >=3.9.0,<3.14
|
|
23
23
|
Description-Content-Type: text/markdown
|
|
24
24
|
License-File: LICENSE.md
|
|
25
|
+
Requires-Dist: pandas
|
|
25
26
|
Requires-Dist: pydantic>=2.7
|
|
27
|
+
Requires-Dist: click
|
|
26
28
|
Requires-Dist: dataclasses_json
|
|
27
29
|
Requires-Dist: tqdm
|
|
28
|
-
Requires-Dist: click
|
|
29
|
-
Requires-Dist: pandas
|
|
30
|
-
Requires-Dist: python-dateutil
|
|
31
30
|
Requires-Dist: opentelemetry-sdk
|
|
31
|
+
Requires-Dist: python-dateutil
|
|
32
32
|
Provides-Extra: remote
|
|
33
33
|
Requires-Dist: unstructured-client>=0.26.1; extra == "remote"
|
|
34
34
|
Provides-Extra: csv
|
|
@@ -66,41 +66,41 @@ Requires-Dist: pyairtable; extra == "airtable"
|
|
|
66
66
|
Provides-Extra: astradb
|
|
67
67
|
Requires-Dist: astrapy; extra == "astradb"
|
|
68
68
|
Provides-Extra: azure
|
|
69
|
-
Requires-Dist: fsspec; extra == "azure"
|
|
70
69
|
Requires-Dist: adlfs; extra == "azure"
|
|
70
|
+
Requires-Dist: fsspec; extra == "azure"
|
|
71
71
|
Provides-Extra: azure-ai-search
|
|
72
72
|
Requires-Dist: azure-search-documents; extra == "azure-ai-search"
|
|
73
73
|
Provides-Extra: biomed
|
|
74
74
|
Requires-Dist: requests; extra == "biomed"
|
|
75
75
|
Requires-Dist: bs4; extra == "biomed"
|
|
76
76
|
Provides-Extra: box
|
|
77
|
-
Requires-Dist: boxfs; extra == "box"
|
|
78
77
|
Requires-Dist: fsspec; extra == "box"
|
|
78
|
+
Requires-Dist: boxfs; extra == "box"
|
|
79
79
|
Provides-Extra: chroma
|
|
80
80
|
Requires-Dist: chromadb; extra == "chroma"
|
|
81
81
|
Provides-Extra: clarifai
|
|
82
82
|
Requires-Dist: clarifai; extra == "clarifai"
|
|
83
83
|
Provides-Extra: confluence
|
|
84
|
-
Requires-Dist: atlassian-python-api; extra == "confluence"
|
|
85
84
|
Requires-Dist: requests; extra == "confluence"
|
|
85
|
+
Requires-Dist: atlassian-python-api; extra == "confluence"
|
|
86
86
|
Provides-Extra: couchbase
|
|
87
87
|
Requires-Dist: couchbase; extra == "couchbase"
|
|
88
88
|
Provides-Extra: delta-table
|
|
89
|
-
Requires-Dist: deltalake; extra == "delta-table"
|
|
90
89
|
Requires-Dist: boto3; extra == "delta-table"
|
|
90
|
+
Requires-Dist: deltalake; extra == "delta-table"
|
|
91
91
|
Provides-Extra: discord
|
|
92
92
|
Requires-Dist: discord.py; extra == "discord"
|
|
93
93
|
Provides-Extra: dropbox
|
|
94
|
-
Requires-Dist: fsspec; extra == "dropbox"
|
|
95
94
|
Requires-Dist: dropboxdrivefs; extra == "dropbox"
|
|
95
|
+
Requires-Dist: fsspec; extra == "dropbox"
|
|
96
96
|
Provides-Extra: duckdb
|
|
97
97
|
Requires-Dist: duckdb; extra == "duckdb"
|
|
98
98
|
Provides-Extra: elasticsearch
|
|
99
99
|
Requires-Dist: elasticsearch[async]; extra == "elasticsearch"
|
|
100
100
|
Provides-Extra: gcs
|
|
101
|
-
Requires-Dist: bs4; extra == "gcs"
|
|
102
101
|
Requires-Dist: gcsfs; extra == "gcs"
|
|
103
102
|
Requires-Dist: fsspec; extra == "gcs"
|
|
103
|
+
Requires-Dist: bs4; extra == "gcs"
|
|
104
104
|
Provides-Extra: github
|
|
105
105
|
Requires-Dist: pygithub>1.58.0; extra == "github"
|
|
106
106
|
Requires-Dist: requests; extra == "github"
|
|
@@ -124,23 +124,23 @@ Requires-Dist: pymilvus; extra == "milvus"
|
|
|
124
124
|
Provides-Extra: mongodb
|
|
125
125
|
Requires-Dist: pymongo; extra == "mongodb"
|
|
126
126
|
Provides-Extra: neo4j
|
|
127
|
+
Requires-Dist: cymple; extra == "neo4j"
|
|
127
128
|
Requires-Dist: neo4j-rust-ext; extra == "neo4j"
|
|
128
129
|
Requires-Dist: networkx; extra == "neo4j"
|
|
129
|
-
Requires-Dist: cymple; extra == "neo4j"
|
|
130
130
|
Provides-Extra: notion
|
|
131
|
+
Requires-Dist: notion-client; extra == "notion"
|
|
132
|
+
Requires-Dist: htmlBuilder; extra == "notion"
|
|
131
133
|
Requires-Dist: backoff; extra == "notion"
|
|
132
134
|
Requires-Dist: httpx; extra == "notion"
|
|
133
|
-
Requires-Dist: htmlBuilder; extra == "notion"
|
|
134
|
-
Requires-Dist: notion-client; extra == "notion"
|
|
135
135
|
Provides-Extra: onedrive
|
|
136
|
-
Requires-Dist: Office365-REST-Python-Client; extra == "onedrive"
|
|
137
|
-
Requires-Dist: bs4; extra == "onedrive"
|
|
138
136
|
Requires-Dist: msal; extra == "onedrive"
|
|
137
|
+
Requires-Dist: bs4; extra == "onedrive"
|
|
138
|
+
Requires-Dist: Office365-REST-Python-Client; extra == "onedrive"
|
|
139
139
|
Provides-Extra: opensearch
|
|
140
140
|
Requires-Dist: opensearch-py; extra == "opensearch"
|
|
141
141
|
Provides-Extra: outlook
|
|
142
|
-
Requires-Dist: Office365-REST-Python-Client; extra == "outlook"
|
|
143
142
|
Requires-Dist: msal; extra == "outlook"
|
|
143
|
+
Requires-Dist: Office365-REST-Python-Client; extra == "outlook"
|
|
144
144
|
Provides-Extra: pinecone
|
|
145
145
|
Requires-Dist: pinecone-client>=3.7.1; extra == "pinecone"
|
|
146
146
|
Provides-Extra: postgres
|
|
@@ -152,11 +152,11 @@ Requires-Dist: praw; extra == "reddit"
|
|
|
152
152
|
Provides-Extra: redis
|
|
153
153
|
Requires-Dist: redis; extra == "redis"
|
|
154
154
|
Provides-Extra: s3
|
|
155
|
-
Requires-Dist: fsspec; extra == "s3"
|
|
156
155
|
Requires-Dist: s3fs; extra == "s3"
|
|
156
|
+
Requires-Dist: fsspec; extra == "s3"
|
|
157
157
|
Provides-Extra: sharepoint
|
|
158
|
-
Requires-Dist: Office365-REST-Python-Client; extra == "sharepoint"
|
|
159
158
|
Requires-Dist: msal; extra == "sharepoint"
|
|
159
|
+
Requires-Dist: Office365-REST-Python-Client; extra == "sharepoint"
|
|
160
160
|
Provides-Extra: salesforce
|
|
161
161
|
Requires-Dist: simple-salesforce; extra == "salesforce"
|
|
162
162
|
Provides-Extra: sftp
|
|
@@ -178,18 +178,18 @@ Requires-Dist: databricks-sql-connector; extra == "databricks-delta-tables"
|
|
|
178
178
|
Provides-Extra: singlestore
|
|
179
179
|
Requires-Dist: singlestoredb; extra == "singlestore"
|
|
180
180
|
Provides-Extra: vectara
|
|
181
|
+
Requires-Dist: aiofiles; extra == "vectara"
|
|
181
182
|
Requires-Dist: requests; extra == "vectara"
|
|
182
183
|
Requires-Dist: httpx; extra == "vectara"
|
|
183
|
-
Requires-Dist: aiofiles; extra == "vectara"
|
|
184
184
|
Provides-Extra: vastdb
|
|
185
|
-
Requires-Dist: ibis; extra == "vastdb"
|
|
186
185
|
Requires-Dist: vastdb; extra == "vastdb"
|
|
186
|
+
Requires-Dist: ibis; extra == "vastdb"
|
|
187
187
|
Requires-Dist: pyarrow; extra == "vastdb"
|
|
188
188
|
Provides-Extra: embed-huggingface
|
|
189
189
|
Requires-Dist: sentence-transformers; extra == "embed-huggingface"
|
|
190
190
|
Provides-Extra: embed-octoai
|
|
191
|
-
Requires-Dist: tiktoken; extra == "embed-octoai"
|
|
192
191
|
Requires-Dist: openai; extra == "embed-octoai"
|
|
192
|
+
Requires-Dist: tiktoken; extra == "embed-octoai"
|
|
193
193
|
Provides-Extra: embed-vertexai
|
|
194
194
|
Requires-Dist: vertexai; extra == "embed-vertexai"
|
|
195
195
|
Provides-Extra: embed-voyageai
|
|
@@ -197,11 +197,11 @@ Requires-Dist: voyageai; extra == "embed-voyageai"
|
|
|
197
197
|
Provides-Extra: embed-mixedbreadai
|
|
198
198
|
Requires-Dist: mixedbread-ai; extra == "embed-mixedbreadai"
|
|
199
199
|
Provides-Extra: openai
|
|
200
|
-
Requires-Dist: tiktoken; extra == "openai"
|
|
201
200
|
Requires-Dist: openai; extra == "openai"
|
|
201
|
+
Requires-Dist: tiktoken; extra == "openai"
|
|
202
202
|
Provides-Extra: bedrock
|
|
203
|
-
Requires-Dist: aioboto3; extra == "bedrock"
|
|
204
203
|
Requires-Dist: boto3; extra == "bedrock"
|
|
204
|
+
Requires-Dist: aioboto3; extra == "bedrock"
|
|
205
205
|
Provides-Extra: togetherai
|
|
206
206
|
Requires-Dist: together; extra == "togetherai"
|
|
207
207
|
Dynamic: author
|
|
@@ -5,7 +5,7 @@ test/integration/chunkers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJW
|
|
|
5
5
|
test/integration/chunkers/test_chunkers.py,sha256=USkltQN_mVVCxI0FkJsrS1gnLXlVr-fvsc0tPaK2sWI,1062
|
|
6
6
|
test/integration/connectors/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
7
7
|
test/integration/connectors/conftest.py,sha256=vYs4WDlCuieAwwErkJxCk4a1lGvr3qpeiAm-YaDznSo,1018
|
|
8
|
-
test/integration/connectors/test_astradb.py,sha256=
|
|
8
|
+
test/integration/connectors/test_astradb.py,sha256=pZmUItFzS91etJONk5HaX8ayarXmFH7RhKmtBxmCClQ,8995
|
|
9
9
|
test/integration/connectors/test_azure_ai_search.py,sha256=MxFwk84vI_HT4taQTGrNpJ8ewGPqHSGrx626j8hC_Pw,9695
|
|
10
10
|
test/integration/connectors/test_chroma.py,sha256=NuQv0PWPM0_LQfdPeUd6IYKqaKKXWmVaHGWjq5aBfOY,3721
|
|
11
11
|
test/integration/connectors/test_confluence.py,sha256=Ju0gRQbD2g9l9iRf2HDZKi7RyPnBGtFRWcGpsqhO3F8,3588
|
|
@@ -107,7 +107,7 @@ test/unit/v2/partitioners/test_partitioner.py,sha256=iIYg7IpftV3LusoO4H8tr1IHY1U
|
|
|
107
107
|
test/unit/v2/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
108
108
|
test/unit/v2/utils/data_generator.py,sha256=UoYVNjG4S4wlaA9gceQ82HIpF9_6I1UTHD1_GrQBHp0,973
|
|
109
109
|
unstructured_ingest/__init__.py,sha256=U4S_2y3zgLZVfMenHRaJFBW8yqh2mUBuI291LGQVOJ8,35
|
|
110
|
-
unstructured_ingest/__version__.py,sha256=
|
|
110
|
+
unstructured_ingest/__version__.py,sha256=Shgafr3Iliv3VjkCZFY-nW2PV7lNrzP2f2kMUaHsecA,42
|
|
111
111
|
unstructured_ingest/error.py,sha256=qDncnJgbf5ils956RcO2CGlAKYDT5OaEM9Clv1JVTNc,1448
|
|
112
112
|
unstructured_ingest/interfaces.py,sha256=7DOnDpGvUNlCoFR7UPRGmOarqH5sFtuUOO5vf8X3oTM,31489
|
|
113
113
|
unstructured_ingest/logger.py,sha256=S5nSqGcABoQyeicgRnBQFjDScCaTvFVivOCvbo-laL0,4479
|
|
@@ -424,10 +424,10 @@ unstructured_ingest/v2/processes/partitioner.py,sha256=ZC9mt85I3o_SLR4DvE7vPBGph
|
|
|
424
424
|
unstructured_ingest/v2/processes/uncompress.py,sha256=Z_XfsITGdyaRwhtNUc7bMj5Y2jLuBge8KoK4nxhqKag,2425
|
|
425
425
|
unstructured_ingest/v2/processes/connectors/__init__.py,sha256=KO1zn-96Qa49TOSZn-gv_RUMGMCmUcdtHoeJqCpxPLY,6219
|
|
426
426
|
unstructured_ingest/v2/processes/connectors/airtable.py,sha256=eeZJe-bBNxt5Sa-XEFCdcGeJCguJU5WN2Mv9kLp5dVQ,8917
|
|
427
|
-
unstructured_ingest/v2/processes/connectors/astradb.py,sha256=
|
|
427
|
+
unstructured_ingest/v2/processes/connectors/astradb.py,sha256=3WFJUNEjeuZFhsLW9KzOIOsiStCjpnqKokS1oIQLUR0,17816
|
|
428
428
|
unstructured_ingest/v2/processes/connectors/azure_ai_search.py,sha256=ngPDpU0oZ6m5sxIlB6u5ebQpqCS_SJ-_amCC1KQ03EQ,11529
|
|
429
429
|
unstructured_ingest/v2/processes/connectors/chroma.py,sha256=VHCnM56qNXuHzovJihrNfJnZbWLJShOe8j12PJFrbL0,7219
|
|
430
|
-
unstructured_ingest/v2/processes/connectors/confluence.py,sha256=
|
|
430
|
+
unstructured_ingest/v2/processes/connectors/confluence.py,sha256=uLpbOtTwbl9TmkWVKbAhH-1UOQvYuCN-v1PIA3BFndc,11139
|
|
431
431
|
unstructured_ingest/v2/processes/connectors/couchbase.py,sha256=i7vuNKsUkN93JRVmg4--MO0ZgbjvhIqt46oYqk9zFSQ,12250
|
|
432
432
|
unstructured_ingest/v2/processes/connectors/delta_table.py,sha256=SotSXZQ85_6TO906YvFi3yTml8jE9A_zV6nBJ4oTx8A,7075
|
|
433
433
|
unstructured_ingest/v2/processes/connectors/discord.py,sha256=-e4-cBK4TnHkknK1qIb86AIVMy81lBgC288_iLpTzM8,5246
|
|
@@ -567,9 +567,9 @@ unstructured_ingest/v2/processes/connectors/weaviate/cloud.py,sha256=bXtfEYLquR-
|
|
|
567
567
|
unstructured_ingest/v2/processes/connectors/weaviate/embedded.py,sha256=S8Zg8StuZT-k7tCg1D5YShO1-vJYYk9-M1bE1fIqx64,3014
|
|
568
568
|
unstructured_ingest/v2/processes/connectors/weaviate/local.py,sha256=LuTBKPseVewsz8VqxRPRLfGEm3BeI9nBZxpy7ZU5tOA,2201
|
|
569
569
|
unstructured_ingest/v2/processes/connectors/weaviate/weaviate.py,sha256=UZ_s8dnVNx9BWFG2fPah4VbQbgEDF4nP78bQeU3jg08,12821
|
|
570
|
-
unstructured_ingest-0.5.
|
|
571
|
-
unstructured_ingest-0.5.
|
|
572
|
-
unstructured_ingest-0.5.
|
|
573
|
-
unstructured_ingest-0.5.
|
|
574
|
-
unstructured_ingest-0.5.
|
|
575
|
-
unstructured_ingest-0.5.
|
|
570
|
+
unstructured_ingest-0.5.9.dist-info/LICENSE.md,sha256=SxkKP_62uIAKb9mb1eH7FH4Kn2aYT09fgjKpJt5PyTk,11360
|
|
571
|
+
unstructured_ingest-0.5.9.dist-info/METADATA,sha256=MvJkJj8xsL18KTeSJbMCEyDOXQ9aJ1xh9WYAnnLxizM,8316
|
|
572
|
+
unstructured_ingest-0.5.9.dist-info/WHEEL,sha256=In9FTNxeP60KnTkGw7wk6mJPYd_dQSjEZmXdBdMCI-8,91
|
|
573
|
+
unstructured_ingest-0.5.9.dist-info/entry_points.txt,sha256=gUAAFnjFPnBgThJSEbw0N5ZjxtaKlT1s9e05_arQrNw,70
|
|
574
|
+
unstructured_ingest-0.5.9.dist-info/top_level.txt,sha256=DMuDMHZRMdeay8v8Kdi855muIv92F0OkutvBCaBEW6M,25
|
|
575
|
+
unstructured_ingest-0.5.9.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
{unstructured_ingest-0.5.8.dist-info → unstructured_ingest-0.5.9.dist-info}/entry_points.txt
RENAMED
|
File without changes
|
|
File without changes
|