groundx 2.0.15__py3-none-any.whl → 2.0.17__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- groundx/__init__.py +4 -3
- groundx/core/client_wrapper.py +1 -1
- groundx/documents/__init__.py +0 -3
- groundx/documents/client.py +40 -42
- groundx/types/__init__.py +4 -0
- groundx/{documents/types/website_crawl_request_websites_item.py → types/crawl_website_source.py} +4 -4
- groundx/{documents/types/documents_ingest_local_request_files_item.py → types/ingest_local_document.py} +5 -5
- {groundx-2.0.15.dist-info → groundx-2.0.17.dist-info}/METADATA +1 -1
- {groundx-2.0.15.dist-info → groundx-2.0.17.dist-info}/RECORD +11 -12
- groundx/documents/types/__init__.py +0 -6
- {groundx-2.0.15.dist-info → groundx-2.0.17.dist-info}/LICENSE +0 -0
- {groundx-2.0.15.dist-info → groundx-2.0.17.dist-info}/WHEEL +0 -0
groundx/__init__.py
CHANGED
@@ -7,6 +7,7 @@ from .types import (
|
|
7
7
|
BucketResponse,
|
8
8
|
BucketUpdateDetail,
|
9
9
|
BucketUpdateResponse,
|
10
|
+
CrawlWebsiteSource,
|
10
11
|
CustomerDetail,
|
11
12
|
CustomerResponse,
|
12
13
|
DocumentDetail,
|
@@ -21,6 +22,7 @@ from .types import (
|
|
21
22
|
HealthResponseHealth,
|
22
23
|
HealthService,
|
23
24
|
HealthServiceStatus,
|
25
|
+
IngestLocalDocument,
|
24
26
|
IngestRemoteDocument,
|
25
27
|
IngestResponse,
|
26
28
|
IngestResponseIngest,
|
@@ -45,7 +47,6 @@ from .types import (
|
|
45
47
|
from .errors import BadRequestError, UnauthorizedError
|
46
48
|
from . import buckets, customer, documents, groups, health, search
|
47
49
|
from .client import AsyncGroundX, GroundX
|
48
|
-
from .documents import DocumentsIngestLocalRequestFilesItem, WebsiteCrawlRequestWebsitesItem
|
49
50
|
from .environment import GroundXEnvironment
|
50
51
|
from .search import SearchContentRequestId
|
51
52
|
from .version import __version__
|
@@ -59,6 +60,7 @@ __all__ = [
|
|
59
60
|
"BucketResponse",
|
60
61
|
"BucketUpdateDetail",
|
61
62
|
"BucketUpdateResponse",
|
63
|
+
"CrawlWebsiteSource",
|
62
64
|
"CustomerDetail",
|
63
65
|
"CustomerResponse",
|
64
66
|
"DocumentDetail",
|
@@ -66,7 +68,6 @@ __all__ = [
|
|
66
68
|
"DocumentLookupResponse",
|
67
69
|
"DocumentResponse",
|
68
70
|
"DocumentType",
|
69
|
-
"DocumentsIngestLocalRequestFilesItem",
|
70
71
|
"GroundX",
|
71
72
|
"GroundXEnvironment",
|
72
73
|
"GroupDetail",
|
@@ -76,6 +77,7 @@ __all__ = [
|
|
76
77
|
"HealthResponseHealth",
|
77
78
|
"HealthService",
|
78
79
|
"HealthServiceStatus",
|
80
|
+
"IngestLocalDocument",
|
79
81
|
"IngestRemoteDocument",
|
80
82
|
"IngestResponse",
|
81
83
|
"IngestResponseIngest",
|
@@ -98,7 +100,6 @@ __all__ = [
|
|
98
100
|
"SubscriptionDetail",
|
99
101
|
"SubscriptionDetailMeters",
|
100
102
|
"UnauthorizedError",
|
101
|
-
"WebsiteCrawlRequestWebsitesItem",
|
102
103
|
"__version__",
|
103
104
|
"buckets",
|
104
105
|
"customer",
|
groundx/core/client_wrapper.py
CHANGED
groundx/documents/__init__.py
CHANGED
groundx/documents/client.py
CHANGED
@@ -11,8 +11,8 @@ from ..errors.bad_request_error import BadRequestError
|
|
11
11
|
from ..errors.unauthorized_error import UnauthorizedError
|
12
12
|
from json.decoder import JSONDecodeError
|
13
13
|
from ..core.api_error import ApiError
|
14
|
-
from
|
15
|
-
from
|
14
|
+
from ..types.ingest_local_document import IngestLocalDocument
|
15
|
+
from ..types.crawl_website_source import CrawlWebsiteSource
|
16
16
|
from ..types.process_status_response import ProcessStatusResponse
|
17
17
|
from ..core.jsonable_encoder import jsonable_encoder
|
18
18
|
from ..types.sort import Sort
|
@@ -124,7 +124,7 @@ class DocumentsClient:
|
|
124
124
|
def ingest_local(
|
125
125
|
self,
|
126
126
|
*,
|
127
|
-
|
127
|
+
documents: typing.Optional[typing.List[IngestLocalDocument]] = OMIT,
|
128
128
|
request_options: typing.Optional[RequestOptions] = None,
|
129
129
|
) -> IngestResponse:
|
130
130
|
"""
|
@@ -134,7 +134,7 @@ class DocumentsClient:
|
|
134
134
|
|
135
135
|
Parameters
|
136
136
|
----------
|
137
|
-
|
137
|
+
documents : typing.Optional[typing.List[IngestLocalDocument]]
|
138
138
|
|
139
139
|
request_options : typing.Optional[RequestOptions]
|
140
140
|
Request-specific configuration.
|
@@ -146,19 +146,19 @@ class DocumentsClient:
|
|
146
146
|
|
147
147
|
Examples
|
148
148
|
--------
|
149
|
-
from groundx import GroundX
|
150
|
-
from groundx.documents import DocumentsIngestLocalRequestFilesItem
|
149
|
+
from groundx import GroundX, IngestLocalDocument
|
151
150
|
|
152
151
|
client = GroundX(
|
153
152
|
api_key="YOUR_API_KEY",
|
154
153
|
)
|
155
154
|
client.documents.ingest_local(
|
156
|
-
|
157
|
-
|
158
|
-
bucket_id=
|
159
|
-
file_data="
|
160
|
-
file_name="
|
155
|
+
documents=[
|
156
|
+
IngestLocalDocument(
|
157
|
+
bucket_id=1234,
|
158
|
+
file_data="binary data",
|
159
|
+
file_name="my_file.txt",
|
161
160
|
file_type="txt",
|
161
|
+
search_data={"key": "value"},
|
162
162
|
)
|
163
163
|
],
|
164
164
|
)
|
@@ -167,7 +167,7 @@ class DocumentsClient:
|
|
167
167
|
"v1/ingest/documents/local",
|
168
168
|
method="POST",
|
169
169
|
data={
|
170
|
-
"
|
170
|
+
"documents": documents,
|
171
171
|
},
|
172
172
|
files={},
|
173
173
|
request_options=request_options,
|
@@ -208,10 +208,7 @@ class DocumentsClient:
|
|
208
208
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
209
209
|
|
210
210
|
def crawl_website(
|
211
|
-
self,
|
212
|
-
*,
|
213
|
-
websites: typing.Sequence[WebsiteCrawlRequestWebsitesItem],
|
214
|
-
request_options: typing.Optional[RequestOptions] = None,
|
211
|
+
self, *, websites: typing.Sequence[CrawlWebsiteSource], request_options: typing.Optional[RequestOptions] = None
|
215
212
|
) -> IngestResponse:
|
216
213
|
"""
|
217
214
|
Upload the content of a publicly accessible website for ingestion into a GroundX bucket. This is done by following links within a specified URL, recursively, up to a specified depth or number of pages.
|
@@ -220,7 +217,7 @@ class DocumentsClient:
|
|
220
217
|
|
221
218
|
Parameters
|
222
219
|
----------
|
223
|
-
websites : typing.Sequence[
|
220
|
+
websites : typing.Sequence[CrawlWebsiteSource]
|
224
221
|
|
225
222
|
request_options : typing.Optional[RequestOptions]
|
226
223
|
Request-specific configuration.
|
@@ -232,16 +229,18 @@ class DocumentsClient:
|
|
232
229
|
|
233
230
|
Examples
|
234
231
|
--------
|
235
|
-
from groundx import GroundX
|
236
|
-
from groundx.documents import WebsiteCrawlRequestWebsitesItem
|
232
|
+
from groundx import CrawlWebsiteSource, GroundX
|
237
233
|
|
238
234
|
client = GroundX(
|
239
235
|
api_key="YOUR_API_KEY",
|
240
236
|
)
|
241
237
|
client.documents.crawl_website(
|
242
238
|
websites=[
|
243
|
-
|
244
|
-
bucket_id=
|
239
|
+
CrawlWebsiteSource(
|
240
|
+
bucket_id=1234,
|
241
|
+
cap=100,
|
242
|
+
depth=3,
|
243
|
+
search_data={"key": "value"},
|
245
244
|
source_url="https://my.website.com",
|
246
245
|
)
|
247
246
|
],
|
@@ -252,7 +251,7 @@ class DocumentsClient:
|
|
252
251
|
method="POST",
|
253
252
|
json={
|
254
253
|
"websites": convert_and_respect_annotation_metadata(
|
255
|
-
object_=websites, annotation=typing.Sequence[
|
254
|
+
object_=websites, annotation=typing.Sequence[CrawlWebsiteSource], direction="write"
|
256
255
|
),
|
257
256
|
},
|
258
257
|
headers={
|
@@ -873,7 +872,7 @@ class AsyncDocumentsClient:
|
|
873
872
|
async def ingest_local(
|
874
873
|
self,
|
875
874
|
*,
|
876
|
-
|
875
|
+
documents: typing.Optional[typing.List[IngestLocalDocument]] = OMIT,
|
877
876
|
request_options: typing.Optional[RequestOptions] = None,
|
878
877
|
) -> IngestResponse:
|
879
878
|
"""
|
@@ -883,7 +882,7 @@ class AsyncDocumentsClient:
|
|
883
882
|
|
884
883
|
Parameters
|
885
884
|
----------
|
886
|
-
|
885
|
+
documents : typing.Optional[typing.List[IngestLocalDocument]]
|
887
886
|
|
888
887
|
request_options : typing.Optional[RequestOptions]
|
889
888
|
Request-specific configuration.
|
@@ -897,8 +896,7 @@ class AsyncDocumentsClient:
|
|
897
896
|
--------
|
898
897
|
import asyncio
|
899
898
|
|
900
|
-
from groundx import AsyncGroundX
|
901
|
-
from groundx.documents import DocumentsIngestLocalRequestFilesItem
|
899
|
+
from groundx import AsyncGroundX, IngestLocalDocument
|
902
900
|
|
903
901
|
client = AsyncGroundX(
|
904
902
|
api_key="YOUR_API_KEY",
|
@@ -907,12 +905,13 @@ class AsyncDocumentsClient:
|
|
907
905
|
|
908
906
|
async def main() -> None:
|
909
907
|
await client.documents.ingest_local(
|
910
|
-
|
911
|
-
|
912
|
-
bucket_id=
|
913
|
-
file_data="
|
914
|
-
file_name="
|
908
|
+
documents=[
|
909
|
+
IngestLocalDocument(
|
910
|
+
bucket_id=1234,
|
911
|
+
file_data="binary data",
|
912
|
+
file_name="my_file.txt",
|
915
913
|
file_type="txt",
|
914
|
+
search_data={"key": "value"},
|
916
915
|
)
|
917
916
|
],
|
918
917
|
)
|
@@ -924,7 +923,7 @@ class AsyncDocumentsClient:
|
|
924
923
|
"v1/ingest/documents/local",
|
925
924
|
method="POST",
|
926
925
|
data={
|
927
|
-
"
|
926
|
+
"documents": documents,
|
928
927
|
},
|
929
928
|
files={},
|
930
929
|
request_options=request_options,
|
@@ -965,10 +964,7 @@ class AsyncDocumentsClient:
|
|
965
964
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
966
965
|
|
967
966
|
async def crawl_website(
|
968
|
-
self,
|
969
|
-
*,
|
970
|
-
websites: typing.Sequence[WebsiteCrawlRequestWebsitesItem],
|
971
|
-
request_options: typing.Optional[RequestOptions] = None,
|
967
|
+
self, *, websites: typing.Sequence[CrawlWebsiteSource], request_options: typing.Optional[RequestOptions] = None
|
972
968
|
) -> IngestResponse:
|
973
969
|
"""
|
974
970
|
Upload the content of a publicly accessible website for ingestion into a GroundX bucket. This is done by following links within a specified URL, recursively, up to a specified depth or number of pages.
|
@@ -977,7 +973,7 @@ class AsyncDocumentsClient:
|
|
977
973
|
|
978
974
|
Parameters
|
979
975
|
----------
|
980
|
-
websites : typing.Sequence[
|
976
|
+
websites : typing.Sequence[CrawlWebsiteSource]
|
981
977
|
|
982
978
|
request_options : typing.Optional[RequestOptions]
|
983
979
|
Request-specific configuration.
|
@@ -991,8 +987,7 @@ class AsyncDocumentsClient:
|
|
991
987
|
--------
|
992
988
|
import asyncio
|
993
989
|
|
994
|
-
from groundx import AsyncGroundX
|
995
|
-
from groundx.documents import WebsiteCrawlRequestWebsitesItem
|
990
|
+
from groundx import AsyncGroundX, CrawlWebsiteSource
|
996
991
|
|
997
992
|
client = AsyncGroundX(
|
998
993
|
api_key="YOUR_API_KEY",
|
@@ -1002,8 +997,11 @@ class AsyncDocumentsClient:
|
|
1002
997
|
async def main() -> None:
|
1003
998
|
await client.documents.crawl_website(
|
1004
999
|
websites=[
|
1005
|
-
|
1006
|
-
bucket_id=
|
1000
|
+
CrawlWebsiteSource(
|
1001
|
+
bucket_id=1234,
|
1002
|
+
cap=100,
|
1003
|
+
depth=3,
|
1004
|
+
search_data={"key": "value"},
|
1007
1005
|
source_url="https://my.website.com",
|
1008
1006
|
)
|
1009
1007
|
],
|
@@ -1017,7 +1015,7 @@ class AsyncDocumentsClient:
|
|
1017
1015
|
method="POST",
|
1018
1016
|
json={
|
1019
1017
|
"websites": convert_and_respect_annotation_metadata(
|
1020
|
-
object_=websites, annotation=typing.Sequence[
|
1018
|
+
object_=websites, annotation=typing.Sequence[CrawlWebsiteSource], direction="write"
|
1021
1019
|
),
|
1022
1020
|
},
|
1023
1021
|
headers={
|
groundx/types/__init__.py
CHANGED
@@ -6,6 +6,7 @@ from .bucket_list_response import BucketListResponse
|
|
6
6
|
from .bucket_response import BucketResponse
|
7
7
|
from .bucket_update_detail import BucketUpdateDetail
|
8
8
|
from .bucket_update_response import BucketUpdateResponse
|
9
|
+
from .crawl_website_source import CrawlWebsiteSource
|
9
10
|
from .customer_detail import CustomerDetail
|
10
11
|
from .customer_response import CustomerResponse
|
11
12
|
from .document_detail import DocumentDetail
|
@@ -20,6 +21,7 @@ from .health_response import HealthResponse
|
|
20
21
|
from .health_response_health import HealthResponseHealth
|
21
22
|
from .health_service import HealthService
|
22
23
|
from .health_service_status import HealthServiceStatus
|
24
|
+
from .ingest_local_document import IngestLocalDocument
|
23
25
|
from .ingest_remote_document import IngestRemoteDocument
|
24
26
|
from .ingest_response import IngestResponse
|
25
27
|
from .ingest_response_ingest import IngestResponseIngest
|
@@ -48,6 +50,7 @@ __all__ = [
|
|
48
50
|
"BucketResponse",
|
49
51
|
"BucketUpdateDetail",
|
50
52
|
"BucketUpdateResponse",
|
53
|
+
"CrawlWebsiteSource",
|
51
54
|
"CustomerDetail",
|
52
55
|
"CustomerResponse",
|
53
56
|
"DocumentDetail",
|
@@ -62,6 +65,7 @@ __all__ = [
|
|
62
65
|
"HealthResponseHealth",
|
63
66
|
"HealthService",
|
64
67
|
"HealthServiceStatus",
|
68
|
+
"IngestLocalDocument",
|
65
69
|
"IngestRemoteDocument",
|
66
70
|
"IngestResponse",
|
67
71
|
"IngestResponseIngest",
|
groundx/{documents/types/website_crawl_request_websites_item.py → types/crawl_website_source.py}
RENAMED
@@ -1,14 +1,14 @@
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
2
2
|
|
3
|
-
from
|
3
|
+
from ..core.pydantic_utilities import UniversalBaseModel
|
4
4
|
import typing_extensions
|
5
|
-
from
|
5
|
+
from ..core.serialization import FieldMetadata
|
6
6
|
import pydantic
|
7
7
|
import typing
|
8
|
-
from
|
8
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
9
9
|
|
10
10
|
|
11
|
-
class
|
11
|
+
class CrawlWebsiteSource(UniversalBaseModel):
|
12
12
|
bucket_id: typing_extensions.Annotated[int, FieldMetadata(alias="bucketId")] = pydantic.Field()
|
13
13
|
"""
|
14
14
|
the bucketId of the bucket which this website will be ingested to.
|
@@ -1,15 +1,15 @@
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
2
2
|
|
3
|
-
from
|
3
|
+
from ..core.pydantic_utilities import UniversalBaseModel
|
4
4
|
import typing_extensions
|
5
|
-
from
|
5
|
+
from ..core.serialization import FieldMetadata
|
6
6
|
import pydantic
|
7
|
-
from
|
7
|
+
from .document_type import DocumentType
|
8
8
|
import typing
|
9
|
-
from
|
9
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
10
10
|
|
11
11
|
|
12
|
-
class
|
12
|
+
class IngestLocalDocument(UniversalBaseModel):
|
13
13
|
bucket_id: typing_extensions.Annotated[int, FieldMetadata(alias="bucketId")] = pydantic.Field()
|
14
14
|
"""
|
15
15
|
the bucketId of the bucket which this local file will be ingested to.
|
@@ -1,10 +1,10 @@
|
|
1
|
-
groundx/__init__.py,sha256=
|
1
|
+
groundx/__init__.py,sha256=22BLAf5omcgumWGCWHDb8o-ha8TnNbkO8LcEvpcYU6A,2851
|
2
2
|
groundx/buckets/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
3
3
|
groundx/buckets/client.py,sha256=TofNrkej1AC_-FU5rf_y8KG8ubFUpHtLa8PQ7rqax6E,26537
|
4
4
|
groundx/client.py,sha256=Q1Kw0z6K-z-ShhNyuuPe5fYonM9M2I_55-ukUrUWk1U,6507
|
5
5
|
groundx/core/__init__.py,sha256=SQ85PF84B9MuKnBwHNHWemSGuy-g_515gFYNFhvEE0I,1438
|
6
6
|
groundx/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
7
|
-
groundx/core/client_wrapper.py,sha256=
|
7
|
+
groundx/core/client_wrapper.py,sha256=zY1z1mzje1H39BKU9IR7OxYeBRV-PVvs004iAgfDqxI,1803
|
8
8
|
groundx/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
|
9
9
|
groundx/core/file.py,sha256=d4NNbX8XvXP32z8KpK2Xovv33nFfruIrpz0QWxlgpZk,2663
|
10
10
|
groundx/core/http_client.py,sha256=siUQ6UV0ARZALlxubqWSSAAPC9B4VW8y6MGlHStfaeo,19552
|
@@ -16,11 +16,8 @@ groundx/core/request_options.py,sha256=h0QUNCFVdCW_7GclVySCAY2w4NhtXVBUCmHgmzaxp
|
|
16
16
|
groundx/core/serialization.py,sha256=D9h_t-RQON3-CHWs1C4ESY9B-Yd5d-l5lnTLb_X896g,9601
|
17
17
|
groundx/customer/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
18
18
|
groundx/customer/client.py,sha256=C_JANeDewRD1Kg-q7LPxdiOSWbYSTOiYlBYZLRYPI44,3467
|
19
|
-
groundx/documents/__init__.py,sha256
|
20
|
-
groundx/documents/client.py,sha256=
|
21
|
-
groundx/documents/types/__init__.py,sha256=fmIT0AisC1K0-EQIa1GDcYr70wpsi7QuXBeyqWiZD14,325
|
22
|
-
groundx/documents/types/documents_ingest_local_request_files_item.py,sha256=EpD7TE1us1DAXdcPvI1li-AGUNpEy_f13bBXidmCAL8,1630
|
23
|
-
groundx/documents/types/website_crawl_request_websites_item.py,sha256=6So4stWecfZYPbiQWg6-FgsfIqV4g2ujFXXgn70evNI,1597
|
19
|
+
groundx/documents/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
20
|
+
groundx/documents/client.py,sha256=_l8K11r-p3qaMd0Qn43JjnSN68stMiaRoTzbbukQxl8,58882
|
24
21
|
groundx/environment.py,sha256=CInm1_DKtZ1mrxutmKb1qqv82P33r_S87hZD3Hc1VB0,159
|
25
22
|
groundx/errors/__init__.py,sha256=-prNYsFd8xxM4va0vR1raZjcd10tllOJKyEWjX_pwdU,214
|
26
23
|
groundx/errors/bad_request_error.py,sha256=_EbO8mWqN9kFZPvIap8qa1lL_EWkRcsZe1HKV9GDWJY,264
|
@@ -34,13 +31,14 @@ groundx/search/__init__.py,sha256=RagVzjShP33mDg9o4N3kGzV0egL1RYNjCpXPE8VzMYE,14
|
|
34
31
|
groundx/search/client.py,sha256=10ifg9GyIwIZF13ULfCXF8iFIydq6H6QRDrGPDjpanw,19756
|
35
32
|
groundx/search/types/__init__.py,sha256=fNFXQloPa1PHHO8VZim6KQNMA9N5EZtfSkissdxtY_c,165
|
36
33
|
groundx/search/types/search_content_request_id.py,sha256=us7mYdzR0qPur_wR5I9BhHaLEzC5nLBRna6-xq4M1ec,128
|
37
|
-
groundx/types/__init__.py,sha256=
|
34
|
+
groundx/types/__init__.py,sha256=yCuwY3aySGDKKYYbxFX_v0kyG_ePu1_M0ZryOwMZJVk,3567
|
38
35
|
groundx/types/bounding_box_detail.py,sha256=51qcen326NTHY2ZqH1cFXut0_MCmk39EbLoDAwotdq4,1832
|
39
36
|
groundx/types/bucket_detail.py,sha256=bQjCvfyWydjItmzNNTvH-iWxNDOggd7R7X1alFZzlEY,1511
|
40
37
|
groundx/types/bucket_list_response.py,sha256=jC0NBsLCYDSwQrBzuW0g3PWFycjtKl8YRkKhic_-1DA,650
|
41
38
|
groundx/types/bucket_response.py,sha256=E8V7H2_TVKdmMsGCBjwzdf2bg4rUjiXFnhXtVGVCqZQ,608
|
42
39
|
groundx/types/bucket_update_detail.py,sha256=B4atQMDSXEdx7otcDbvgsrAHtXNz9swMnOsXRe25coc,738
|
43
40
|
groundx/types/bucket_update_response.py,sha256=h5RJTEpc4WPI_C4sPvsJZo7IxKppnPR-I9VGEQryRlc,633
|
41
|
+
groundx/types/crawl_website_source.py,sha256=K4-IoK__K4OekBwmJhClL2rZD1P4rFPR2-Ck02jP2uw,1581
|
44
42
|
groundx/types/customer_detail.py,sha256=RNm0qXvKx6YvVmkVJZeCNIz7n8_siFMXJ_AGtH3i5Z0,1094
|
45
43
|
groundx/types/customer_response.py,sha256=_RbuanXhCWQkCeQ0dkwPgsjNBoBgNpixiNfRpXcMew8,618
|
46
44
|
groundx/types/document_detail.py,sha256=i1UfcQAGYo9v1HwrrpzQPw_O0qA7IOXwOUuPV1yU8nI,2323
|
@@ -55,6 +53,7 @@ groundx/types/health_response.py,sha256=3UpYL2IZb56tTo-fOpSU-0OTRyWgpYiB3pMU3sfj
|
|
55
53
|
groundx/types/health_response_health.py,sha256=I0QeEljFp6l5LCJbCTArW031Om84egePgnGdtE6WXlI,632
|
56
54
|
groundx/types/health_service.py,sha256=M1-h1EJSpAXw-j3pY-09_g_WKkO0spdj8e7pgPzSGf0,1083
|
57
55
|
groundx/types/health_service_status.py,sha256=ugKJXlx8QGi83n_J6s1frFrW1hYfOn3Dlb_pPNexwMA,185
|
56
|
+
groundx/types/ingest_local_document.py,sha256=2T1HXR2a-BDj5LEOTM98Sl4sgjVMWbVShQLn8MTR6QA,1602
|
58
57
|
groundx/types/ingest_remote_document.py,sha256=xlPA4SYoUgoGXpxZhyORdezxIPGmr4wneav2ZEVmmOY,1683
|
59
58
|
groundx/types/ingest_response.py,sha256=139rn8wpT44jlUzYXiy0r8XzN2U_OtdLltpSbRU0TyA,633
|
60
59
|
groundx/types/ingest_response_ingest.py,sha256=8FKApYNvS6KFxEKm05pKpKJ0BAagxoE0cWeTt-qjm1g,781
|
@@ -76,7 +75,7 @@ groundx/types/sort_order.py,sha256=hfJkStz6zHf3iWQFaVLkNCZPdyj5JS7TsQlN4Ij8Q5A,1
|
|
76
75
|
groundx/types/subscription_detail.py,sha256=WNfUw2EMVECIvNYcV2s51zZ6T3Utc4zYXw63bPaeM6U,768
|
77
76
|
groundx/types/subscription_detail_meters.py,sha256=lBa8-1QlMVHjr5RLGqhiTKnD1KMM0AAHTWvz9TVtG8w,830
|
78
77
|
groundx/version.py,sha256=1yVogKaq260fQfckM2RYN2144SEw0QROsZW8ICtkG4U,74
|
79
|
-
groundx-2.0.
|
80
|
-
groundx-2.0.
|
81
|
-
groundx-2.0.
|
82
|
-
groundx-2.0.
|
78
|
+
groundx-2.0.17.dist-info/LICENSE,sha256=8dMPYAFBTA7O4DUxhrEKEks8CL2waCMYM6dHohW4xrI,1065
|
79
|
+
groundx-2.0.17.dist-info/METADATA,sha256=HZspuVrfswC62NtJa1cUVkS53KO8stJ8xnqTNfFx4lE,5206
|
80
|
+
groundx-2.0.17.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
|
81
|
+
groundx-2.0.17.dist-info/RECORD,,
|
@@ -1,6 +0,0 @@
|
|
1
|
-
# This file was auto-generated by Fern from our API Definition.
|
2
|
-
|
3
|
-
from .documents_ingest_local_request_files_item import DocumentsIngestLocalRequestFilesItem
|
4
|
-
from .website_crawl_request_websites_item import WebsiteCrawlRequestWebsitesItem
|
5
|
-
|
6
|
-
__all__ = ["DocumentsIngestLocalRequestFilesItem", "WebsiteCrawlRequestWebsitesItem"]
|
File without changes
|
File without changes
|