groundx 2.0.15__tar.gz → 2.0.17__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (83) hide show
  1. {groundx-2.0.15 → groundx-2.0.17}/PKG-INFO +1 -1
  2. {groundx-2.0.15 → groundx-2.0.17}/pyproject.toml +1 -1
  3. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/__init__.py +4 -3
  4. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/client_wrapper.py +1 -1
  5. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/documents/client.py +40 -42
  6. groundx-2.0.17/src/groundx/health/__init__.py +2 -0
  7. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/__init__.py +4 -0
  8. groundx-2.0.15/src/groundx/documents/types/website_crawl_request_websites_item.py → groundx-2.0.17/src/groundx/types/crawl_website_source.py +4 -4
  9. groundx-2.0.15/src/groundx/documents/types/documents_ingest_local_request_files_item.py → groundx-2.0.17/src/groundx/types/ingest_local_document.py +5 -5
  10. groundx-2.0.15/src/groundx/documents/__init__.py +0 -5
  11. groundx-2.0.15/src/groundx/documents/types/__init__.py +0 -6
  12. {groundx-2.0.15 → groundx-2.0.17}/LICENSE +0 -0
  13. {groundx-2.0.15 → groundx-2.0.17}/README.md +0 -0
  14. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/buckets/__init__.py +0 -0
  15. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/buckets/client.py +0 -0
  16. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/client.py +0 -0
  17. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/__init__.py +0 -0
  18. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/api_error.py +0 -0
  19. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/datetime_utils.py +0 -0
  20. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/file.py +0 -0
  21. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/http_client.py +0 -0
  22. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/jsonable_encoder.py +0 -0
  23. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/pydantic_utilities.py +0 -0
  24. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/query_encoder.py +0 -0
  25. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/remove_none_from_dict.py +0 -0
  26. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/request_options.py +0 -0
  27. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/core/serialization.py +0 -0
  28. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/customer/__init__.py +0 -0
  29. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/customer/client.py +0 -0
  30. {groundx-2.0.15/src/groundx/groups → groundx-2.0.17/src/groundx/documents}/__init__.py +0 -0
  31. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/environment.py +0 -0
  32. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/errors/__init__.py +0 -0
  33. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/errors/bad_request_error.py +0 -0
  34. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/errors/unauthorized_error.py +0 -0
  35. {groundx-2.0.15/src/groundx/health → groundx-2.0.17/src/groundx/groups}/__init__.py +0 -0
  36. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/groups/client.py +0 -0
  37. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/health/client.py +0 -0
  38. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/py.typed +0 -0
  39. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/search/__init__.py +0 -0
  40. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/search/client.py +0 -0
  41. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/search/types/__init__.py +0 -0
  42. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/search/types/search_content_request_id.py +0 -0
  43. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bounding_box_detail.py +0 -0
  44. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bucket_detail.py +0 -0
  45. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bucket_list_response.py +0 -0
  46. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bucket_response.py +0 -0
  47. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bucket_update_detail.py +0 -0
  48. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/bucket_update_response.py +0 -0
  49. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/customer_detail.py +0 -0
  50. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/customer_response.py +0 -0
  51. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/document_detail.py +0 -0
  52. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/document_list_response.py +0 -0
  53. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/document_lookup_response.py +0 -0
  54. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/document_response.py +0 -0
  55. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/document_type.py +0 -0
  56. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/group_detail.py +0 -0
  57. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/group_list_response.py +0 -0
  58. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/group_response.py +0 -0
  59. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/health_response.py +0 -0
  60. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/health_response_health.py +0 -0
  61. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/health_service.py +0 -0
  62. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/health_service_status.py +0 -0
  63. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/ingest_remote_document.py +0 -0
  64. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/ingest_response.py +0 -0
  65. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/ingest_response_ingest.py +0 -0
  66. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/message_response.py +0 -0
  67. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/meter_detail.py +0 -0
  68. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response.py +0 -0
  69. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest.py +0 -0
  70. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest_progress.py +0 -0
  71. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest_progress_cancelled.py +0 -0
  72. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest_progress_complete.py +0 -0
  73. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest_progress_errors.py +0 -0
  74. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/process_status_response_ingest_progress_processing.py +0 -0
  75. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/processing_status.py +0 -0
  76. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/search_response.py +0 -0
  77. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/search_response_search.py +0 -0
  78. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/search_result_item.py +0 -0
  79. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/sort.py +0 -0
  80. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/sort_order.py +0 -0
  81. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/subscription_detail.py +0 -0
  82. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/types/subscription_detail_meters.py +0 -0
  83. {groundx-2.0.15 → groundx-2.0.17}/src/groundx/version.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: groundx
3
- Version: 2.0.15
3
+ Version: 2.0.17
4
4
  Summary:
5
5
  License: MIT
6
6
  Requires-Python: >=3.8,<4.0
@@ -1,6 +1,6 @@
1
1
  [tool.poetry]
2
2
  name = "groundx"
3
- version = "2.0.15"
3
+ version = "2.0.17"
4
4
  description = ""
5
5
  readme = "README.md"
6
6
  authors = []
@@ -7,6 +7,7 @@ from .types import (
7
7
  BucketResponse,
8
8
  BucketUpdateDetail,
9
9
  BucketUpdateResponse,
10
+ CrawlWebsiteSource,
10
11
  CustomerDetail,
11
12
  CustomerResponse,
12
13
  DocumentDetail,
@@ -21,6 +22,7 @@ from .types import (
21
22
  HealthResponseHealth,
22
23
  HealthService,
23
24
  HealthServiceStatus,
25
+ IngestLocalDocument,
24
26
  IngestRemoteDocument,
25
27
  IngestResponse,
26
28
  IngestResponseIngest,
@@ -45,7 +47,6 @@ from .types import (
45
47
  from .errors import BadRequestError, UnauthorizedError
46
48
  from . import buckets, customer, documents, groups, health, search
47
49
  from .client import AsyncGroundX, GroundX
48
- from .documents import DocumentsIngestLocalRequestFilesItem, WebsiteCrawlRequestWebsitesItem
49
50
  from .environment import GroundXEnvironment
50
51
  from .search import SearchContentRequestId
51
52
  from .version import __version__
@@ -59,6 +60,7 @@ __all__ = [
59
60
  "BucketResponse",
60
61
  "BucketUpdateDetail",
61
62
  "BucketUpdateResponse",
63
+ "CrawlWebsiteSource",
62
64
  "CustomerDetail",
63
65
  "CustomerResponse",
64
66
  "DocumentDetail",
@@ -66,7 +68,6 @@ __all__ = [
66
68
  "DocumentLookupResponse",
67
69
  "DocumentResponse",
68
70
  "DocumentType",
69
- "DocumentsIngestLocalRequestFilesItem",
70
71
  "GroundX",
71
72
  "GroundXEnvironment",
72
73
  "GroupDetail",
@@ -76,6 +77,7 @@ __all__ = [
76
77
  "HealthResponseHealth",
77
78
  "HealthService",
78
79
  "HealthServiceStatus",
80
+ "IngestLocalDocument",
79
81
  "IngestRemoteDocument",
80
82
  "IngestResponse",
81
83
  "IngestResponseIngest",
@@ -98,7 +100,6 @@ __all__ = [
98
100
  "SubscriptionDetail",
99
101
  "SubscriptionDetailMeters",
100
102
  "UnauthorizedError",
101
- "WebsiteCrawlRequestWebsitesItem",
102
103
  "__version__",
103
104
  "buckets",
104
105
  "customer",
@@ -16,7 +16,7 @@ class BaseClientWrapper:
16
16
  headers: typing.Dict[str, str] = {
17
17
  "X-Fern-Language": "Python",
18
18
  "X-Fern-SDK-Name": "groundx",
19
- "X-Fern-SDK-Version": "2.0.15",
19
+ "X-Fern-SDK-Version": "2.0.17",
20
20
  }
21
21
  headers["X-API-Key"] = self.api_key
22
22
  return headers
@@ -11,8 +11,8 @@ from ..errors.bad_request_error import BadRequestError
11
11
  from ..errors.unauthorized_error import UnauthorizedError
12
12
  from json.decoder import JSONDecodeError
13
13
  from ..core.api_error import ApiError
14
- from .types.documents_ingest_local_request_files_item import DocumentsIngestLocalRequestFilesItem
15
- from .types.website_crawl_request_websites_item import WebsiteCrawlRequestWebsitesItem
14
+ from ..types.ingest_local_document import IngestLocalDocument
15
+ from ..types.crawl_website_source import CrawlWebsiteSource
16
16
  from ..types.process_status_response import ProcessStatusResponse
17
17
  from ..core.jsonable_encoder import jsonable_encoder
18
18
  from ..types.sort import Sort
@@ -124,7 +124,7 @@ class DocumentsClient:
124
124
  def ingest_local(
125
125
  self,
126
126
  *,
127
- files: typing.List[DocumentsIngestLocalRequestFilesItem],
127
+ documents: typing.Optional[typing.List[IngestLocalDocument]] = OMIT,
128
128
  request_options: typing.Optional[RequestOptions] = None,
129
129
  ) -> IngestResponse:
130
130
  """
@@ -134,7 +134,7 @@ class DocumentsClient:
134
134
 
135
135
  Parameters
136
136
  ----------
137
- files : typing.List[DocumentsIngestLocalRequestFilesItem]
137
+ documents : typing.Optional[typing.List[IngestLocalDocument]]
138
138
 
139
139
  request_options : typing.Optional[RequestOptions]
140
140
  Request-specific configuration.
@@ -146,19 +146,19 @@ class DocumentsClient:
146
146
 
147
147
  Examples
148
148
  --------
149
- from groundx import GroundX
150
- from groundx.documents import DocumentsIngestLocalRequestFilesItem
149
+ from groundx import GroundX, IngestLocalDocument
151
150
 
152
151
  client = GroundX(
153
152
  api_key="YOUR_API_KEY",
154
153
  )
155
154
  client.documents.ingest_local(
156
- files=[
157
- DocumentsIngestLocalRequestFilesItem(
158
- bucket_id=1,
159
- file_data="fileData",
160
- file_name="fileName",
155
+ documents=[
156
+ IngestLocalDocument(
157
+ bucket_id=1234,
158
+ file_data="binary data",
159
+ file_name="my_file.txt",
161
160
  file_type="txt",
161
+ search_data={"key": "value"},
162
162
  )
163
163
  ],
164
164
  )
@@ -167,7 +167,7 @@ class DocumentsClient:
167
167
  "v1/ingest/documents/local",
168
168
  method="POST",
169
169
  data={
170
- "files": files,
170
+ "documents": documents,
171
171
  },
172
172
  files={},
173
173
  request_options=request_options,
@@ -208,10 +208,7 @@ class DocumentsClient:
208
208
  raise ApiError(status_code=_response.status_code, body=_response_json)
209
209
 
210
210
  def crawl_website(
211
- self,
212
- *,
213
- websites: typing.Sequence[WebsiteCrawlRequestWebsitesItem],
214
- request_options: typing.Optional[RequestOptions] = None,
211
+ self, *, websites: typing.Sequence[CrawlWebsiteSource], request_options: typing.Optional[RequestOptions] = None
215
212
  ) -> IngestResponse:
216
213
  """
217
214
  Upload the content of a publicly accessible website for ingestion into a GroundX bucket. This is done by following links within a specified URL, recursively, up to a specified depth or number of pages.
@@ -220,7 +217,7 @@ class DocumentsClient:
220
217
 
221
218
  Parameters
222
219
  ----------
223
- websites : typing.Sequence[WebsiteCrawlRequestWebsitesItem]
220
+ websites : typing.Sequence[CrawlWebsiteSource]
224
221
 
225
222
  request_options : typing.Optional[RequestOptions]
226
223
  Request-specific configuration.
@@ -232,16 +229,18 @@ class DocumentsClient:
232
229
 
233
230
  Examples
234
231
  --------
235
- from groundx import GroundX
236
- from groundx.documents import WebsiteCrawlRequestWebsitesItem
232
+ from groundx import CrawlWebsiteSource, GroundX
237
233
 
238
234
  client = GroundX(
239
235
  api_key="YOUR_API_KEY",
240
236
  )
241
237
  client.documents.crawl_website(
242
238
  websites=[
243
- WebsiteCrawlRequestWebsitesItem(
244
- bucket_id=123,
239
+ CrawlWebsiteSource(
240
+ bucket_id=1234,
241
+ cap=100,
242
+ depth=3,
243
+ search_data={"key": "value"},
245
244
  source_url="https://my.website.com",
246
245
  )
247
246
  ],
@@ -252,7 +251,7 @@ class DocumentsClient:
252
251
  method="POST",
253
252
  json={
254
253
  "websites": convert_and_respect_annotation_metadata(
255
- object_=websites, annotation=typing.Sequence[WebsiteCrawlRequestWebsitesItem], direction="write"
254
+ object_=websites, annotation=typing.Sequence[CrawlWebsiteSource], direction="write"
256
255
  ),
257
256
  },
258
257
  headers={
@@ -873,7 +872,7 @@ class AsyncDocumentsClient:
873
872
  async def ingest_local(
874
873
  self,
875
874
  *,
876
- files: typing.List[DocumentsIngestLocalRequestFilesItem],
875
+ documents: typing.Optional[typing.List[IngestLocalDocument]] = OMIT,
877
876
  request_options: typing.Optional[RequestOptions] = None,
878
877
  ) -> IngestResponse:
879
878
  """
@@ -883,7 +882,7 @@ class AsyncDocumentsClient:
883
882
 
884
883
  Parameters
885
884
  ----------
886
- files : typing.List[DocumentsIngestLocalRequestFilesItem]
885
+ documents : typing.Optional[typing.List[IngestLocalDocument]]
887
886
 
888
887
  request_options : typing.Optional[RequestOptions]
889
888
  Request-specific configuration.
@@ -897,8 +896,7 @@ class AsyncDocumentsClient:
897
896
  --------
898
897
  import asyncio
899
898
 
900
- from groundx import AsyncGroundX
901
- from groundx.documents import DocumentsIngestLocalRequestFilesItem
899
+ from groundx import AsyncGroundX, IngestLocalDocument
902
900
 
903
901
  client = AsyncGroundX(
904
902
  api_key="YOUR_API_KEY",
@@ -907,12 +905,13 @@ class AsyncDocumentsClient:
907
905
 
908
906
  async def main() -> None:
909
907
  await client.documents.ingest_local(
910
- files=[
911
- DocumentsIngestLocalRequestFilesItem(
912
- bucket_id=1,
913
- file_data="fileData",
914
- file_name="fileName",
908
+ documents=[
909
+ IngestLocalDocument(
910
+ bucket_id=1234,
911
+ file_data="binary data",
912
+ file_name="my_file.txt",
915
913
  file_type="txt",
914
+ search_data={"key": "value"},
916
915
  )
917
916
  ],
918
917
  )
@@ -924,7 +923,7 @@ class AsyncDocumentsClient:
924
923
  "v1/ingest/documents/local",
925
924
  method="POST",
926
925
  data={
927
- "files": files,
926
+ "documents": documents,
928
927
  },
929
928
  files={},
930
929
  request_options=request_options,
@@ -965,10 +964,7 @@ class AsyncDocumentsClient:
965
964
  raise ApiError(status_code=_response.status_code, body=_response_json)
966
965
 
967
966
  async def crawl_website(
968
- self,
969
- *,
970
- websites: typing.Sequence[WebsiteCrawlRequestWebsitesItem],
971
- request_options: typing.Optional[RequestOptions] = None,
967
+ self, *, websites: typing.Sequence[CrawlWebsiteSource], request_options: typing.Optional[RequestOptions] = None
972
968
  ) -> IngestResponse:
973
969
  """
974
970
  Upload the content of a publicly accessible website for ingestion into a GroundX bucket. This is done by following links within a specified URL, recursively, up to a specified depth or number of pages.
@@ -977,7 +973,7 @@ class AsyncDocumentsClient:
977
973
 
978
974
  Parameters
979
975
  ----------
980
- websites : typing.Sequence[WebsiteCrawlRequestWebsitesItem]
976
+ websites : typing.Sequence[CrawlWebsiteSource]
981
977
 
982
978
  request_options : typing.Optional[RequestOptions]
983
979
  Request-specific configuration.
@@ -991,8 +987,7 @@ class AsyncDocumentsClient:
991
987
  --------
992
988
  import asyncio
993
989
 
994
- from groundx import AsyncGroundX
995
- from groundx.documents import WebsiteCrawlRequestWebsitesItem
990
+ from groundx import AsyncGroundX, CrawlWebsiteSource
996
991
 
997
992
  client = AsyncGroundX(
998
993
  api_key="YOUR_API_KEY",
@@ -1002,8 +997,11 @@ class AsyncDocumentsClient:
1002
997
  async def main() -> None:
1003
998
  await client.documents.crawl_website(
1004
999
  websites=[
1005
- WebsiteCrawlRequestWebsitesItem(
1006
- bucket_id=123,
1000
+ CrawlWebsiteSource(
1001
+ bucket_id=1234,
1002
+ cap=100,
1003
+ depth=3,
1004
+ search_data={"key": "value"},
1007
1005
  source_url="https://my.website.com",
1008
1006
  )
1009
1007
  ],
@@ -1017,7 +1015,7 @@ class AsyncDocumentsClient:
1017
1015
  method="POST",
1018
1016
  json={
1019
1017
  "websites": convert_and_respect_annotation_metadata(
1020
- object_=websites, annotation=typing.Sequence[WebsiteCrawlRequestWebsitesItem], direction="write"
1018
+ object_=websites, annotation=typing.Sequence[CrawlWebsiteSource], direction="write"
1021
1019
  ),
1022
1020
  },
1023
1021
  headers={
@@ -0,0 +1,2 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
@@ -6,6 +6,7 @@ from .bucket_list_response import BucketListResponse
6
6
  from .bucket_response import BucketResponse
7
7
  from .bucket_update_detail import BucketUpdateDetail
8
8
  from .bucket_update_response import BucketUpdateResponse
9
+ from .crawl_website_source import CrawlWebsiteSource
9
10
  from .customer_detail import CustomerDetail
10
11
  from .customer_response import CustomerResponse
11
12
  from .document_detail import DocumentDetail
@@ -20,6 +21,7 @@ from .health_response import HealthResponse
20
21
  from .health_response_health import HealthResponseHealth
21
22
  from .health_service import HealthService
22
23
  from .health_service_status import HealthServiceStatus
24
+ from .ingest_local_document import IngestLocalDocument
23
25
  from .ingest_remote_document import IngestRemoteDocument
24
26
  from .ingest_response import IngestResponse
25
27
  from .ingest_response_ingest import IngestResponseIngest
@@ -48,6 +50,7 @@ __all__ = [
48
50
  "BucketResponse",
49
51
  "BucketUpdateDetail",
50
52
  "BucketUpdateResponse",
53
+ "CrawlWebsiteSource",
51
54
  "CustomerDetail",
52
55
  "CustomerResponse",
53
56
  "DocumentDetail",
@@ -62,6 +65,7 @@ __all__ = [
62
65
  "HealthResponseHealth",
63
66
  "HealthService",
64
67
  "HealthServiceStatus",
68
+ "IngestLocalDocument",
65
69
  "IngestRemoteDocument",
66
70
  "IngestResponse",
67
71
  "IngestResponseIngest",
@@ -1,14 +1,14 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
- from ...core.pydantic_utilities import UniversalBaseModel
3
+ from ..core.pydantic_utilities import UniversalBaseModel
4
4
  import typing_extensions
5
- from ...core.serialization import FieldMetadata
5
+ from ..core.serialization import FieldMetadata
6
6
  import pydantic
7
7
  import typing
8
- from ...core.pydantic_utilities import IS_PYDANTIC_V2
8
+ from ..core.pydantic_utilities import IS_PYDANTIC_V2
9
9
 
10
10
 
11
- class WebsiteCrawlRequestWebsitesItem(UniversalBaseModel):
11
+ class CrawlWebsiteSource(UniversalBaseModel):
12
12
  bucket_id: typing_extensions.Annotated[int, FieldMetadata(alias="bucketId")] = pydantic.Field()
13
13
  """
14
14
  the bucketId of the bucket which this website will be ingested to.
@@ -1,15 +1,15 @@
1
1
  # This file was auto-generated by Fern from our API Definition.
2
2
 
3
- from ...core.pydantic_utilities import UniversalBaseModel
3
+ from ..core.pydantic_utilities import UniversalBaseModel
4
4
  import typing_extensions
5
- from ...core.serialization import FieldMetadata
5
+ from ..core.serialization import FieldMetadata
6
6
  import pydantic
7
- from ...types.document_type import DocumentType
7
+ from .document_type import DocumentType
8
8
  import typing
9
- from ...core.pydantic_utilities import IS_PYDANTIC_V2
9
+ from ..core.pydantic_utilities import IS_PYDANTIC_V2
10
10
 
11
11
 
12
- class DocumentsIngestLocalRequestFilesItem(UniversalBaseModel):
12
+ class IngestLocalDocument(UniversalBaseModel):
13
13
  bucket_id: typing_extensions.Annotated[int, FieldMetadata(alias="bucketId")] = pydantic.Field()
14
14
  """
15
15
  the bucketId of the bucket which this local file will be ingested to.
@@ -1,5 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- from .types import DocumentsIngestLocalRequestFilesItem, WebsiteCrawlRequestWebsitesItem
4
-
5
- __all__ = ["DocumentsIngestLocalRequestFilesItem", "WebsiteCrawlRequestWebsitesItem"]
@@ -1,6 +0,0 @@
1
- # This file was auto-generated by Fern from our API Definition.
2
-
3
- from .documents_ingest_local_request_files_item import DocumentsIngestLocalRequestFilesItem
4
- from .website_crawl_request_websites_item import WebsiteCrawlRequestWebsitesItem
5
-
6
- __all__ = ["DocumentsIngestLocalRequestFilesItem", "WebsiteCrawlRequestWebsitesItem"]
File without changes
File without changes
File without changes
File without changes