lockss-pyclient 0.1.0.dev1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- lockss/pyclient/__init__.py +67 -0
- lockss/pyclient/config/__init__.py +42 -0
- lockss/pyclient/config/api/__init__.py +12 -0
- lockss/pyclient/config/api/aus_api.py +2195 -0
- lockss/pyclient/config/api/config_api.py +718 -0
- lockss/pyclient/config/api/plugins_api.py +128 -0
- lockss/pyclient/config/api/status_api.py +120 -0
- lockss/pyclient/config/api/tdb_api.py +318 -0
- lockss/pyclient/config/api/users_api.py +516 -0
- lockss/pyclient/config/api/utils_api.py +128 -0
- lockss/pyclient/config/api_client.py +632 -0
- lockss/pyclient/config/configuration.py +254 -0
- lockss/pyclient/config/models/__init__.py +30 -0
- lockss/pyclient/config/models/api_status.py +344 -0
- lockss/pyclient/config/models/au_configuration.py +142 -0
- lockss/pyclient/config/models/au_status.py +113 -0
- lockss/pyclient/config/models/au_ws_result.py +113 -0
- lockss/pyclient/config/models/auids_body.py +168 -0
- lockss/pyclient/config/models/check_substance_result.py +212 -0
- lockss/pyclient/config/models/content_configuration_result.py +200 -0
- lockss/pyclient/config/models/file_section_name_body.py +113 -0
- lockss/pyclient/config/models/platform_configuration_ws_result.py +113 -0
- lockss/pyclient/config/models/plugin_ws_result.py +345 -0
- lockss/pyclient/config/models/request_au_control_result.py +171 -0
- lockss/pyclient/config/models/tdb_au_ws_result.py +360 -0
- lockss/pyclient/config/models/tdb_publisher_ws_result.py +113 -0
- lockss/pyclient/config/models/tdb_title_ws_result.py +390 -0
- lockss/pyclient/config/rest.py +317 -0
- lockss/pyclient/crawler/__init__.py +45 -0
- lockss/pyclient/crawler/api/__init__.py +10 -0
- lockss/pyclient/crawler/api/crawlers_api.py +215 -0
- lockss/pyclient/crawler/api/crawls_api.py +952 -0
- lockss/pyclient/crawler/api/jobs_api.py +504 -0
- lockss/pyclient/crawler/api/status_api.py +120 -0
- lockss/pyclient/crawler/api/ws_api.py +128 -0
- lockss/pyclient/crawler/api_client.py +632 -0
- lockss/pyclient/crawler/configuration.py +254 -0
- lockss/pyclient/crawler/models/__init__.py +35 -0
- lockss/pyclient/crawler/models/api_status.py +344 -0
- lockss/pyclient/crawler/models/counter.py +142 -0
- lockss/pyclient/crawler/models/crawl_desc.py +344 -0
- lockss/pyclient/crawler/models/crawl_job.py +280 -0
- lockss/pyclient/crawler/models/crawl_pager.py +140 -0
- lockss/pyclient/crawler/models/crawl_status.py +780 -0
- lockss/pyclient/crawler/models/crawl_ws_result.py +814 -0
- lockss/pyclient/crawler/models/crawl_ws_result_pages_with_errors.py +162 -0
- lockss/pyclient/crawler/models/crawler_config.py +142 -0
- lockss/pyclient/crawler/models/crawler_status.py +279 -0
- lockss/pyclient/crawler/models/crawler_statuses.py +112 -0
- lockss/pyclient/crawler/models/error_result.py +164 -0
- lockss/pyclient/crawler/models/job_pager.py +140 -0
- lockss/pyclient/crawler/models/job_status.py +147 -0
- lockss/pyclient/crawler/models/mime_counter.py +169 -0
- lockss/pyclient/crawler/models/page_info.py +228 -0
- lockss/pyclient/crawler/models/url_error.py +148 -0
- lockss/pyclient/crawler/models/url_info.py +167 -0
- lockss/pyclient/crawler/models/url_pager.py +140 -0
- lockss/pyclient/crawler/rest.py +317 -0
- lockss/pyclient/md/__init__.py +36 -0
- lockss/pyclient/md/api/__init__.py +9 -0
- lockss/pyclient/md/api/mdupdates_api.py +508 -0
- lockss/pyclient/md/api/metadata_api.py +136 -0
- lockss/pyclient/md/api/status_api.py +120 -0
- lockss/pyclient/md/api/urls_api.py +224 -0
- lockss/pyclient/md/api_client.py +632 -0
- lockss/pyclient/md/configuration.py +254 -0
- lockss/pyclient/md/models/__init__.py +27 -0
- lockss/pyclient/md/models/api_status.py +344 -0
- lockss/pyclient/md/models/au.py +169 -0
- lockss/pyclient/md/models/au_metadata_page_info.py +140 -0
- lockss/pyclient/md/models/error_result.py +164 -0
- lockss/pyclient/md/models/item_metadata.py +196 -0
- lockss/pyclient/md/models/job.py +280 -0
- lockss/pyclient/md/models/job_page_info.py +140 -0
- lockss/pyclient/md/models/metadata_update_spec.py +142 -0
- lockss/pyclient/md/models/page_info.py +228 -0
- lockss/pyclient/md/models/status.py +142 -0
- lockss/pyclient/md/models/url_info.py +142 -0
- lockss/pyclient/md/rest.py +317 -0
- lockss/pyclient/poller/__init__.py +54 -0
- lockss/pyclient/poller/api/__init__.py +13 -0
- lockss/pyclient/poller/api/export_api.py +156 -0
- lockss/pyclient/poller/api/hash_api.py +413 -0
- lockss/pyclient/poller/api/import_api.py +157 -0
- lockss/pyclient/poller/api/poll_detail_api.py +374 -0
- lockss/pyclient/poller/api/poller_polls_api.py +223 -0
- lockss/pyclient/poller/api/repo_api.py +223 -0
- lockss/pyclient/poller/api/service_api.py +694 -0
- lockss/pyclient/poller/api/voter_polls_api.py +223 -0
- lockss/pyclient/poller/api_client.py +632 -0
- lockss/pyclient/poller/configuration.py +254 -0
- lockss/pyclient/poller/models/__init__.py +41 -0
- lockss/pyclient/poller/models/api_status.py +344 -0
- lockss/pyclient/poller/models/aus_import_body.py +199 -0
- lockss/pyclient/poller/models/cached_uri_set_spec.py +169 -0
- lockss/pyclient/poller/models/error_result.py +164 -0
- lockss/pyclient/poller/models/hasher_ws_params.py +432 -0
- lockss/pyclient/poller/models/link_desc.py +141 -0
- lockss/pyclient/poller/models/page_desc.py +227 -0
- lockss/pyclient/poller/models/peer_data.py +638 -0
- lockss/pyclient/poller/models/peer_ws_result.py +113 -0
- lockss/pyclient/poller/models/poll_desc.py +285 -0
- lockss/pyclient/poller/models/poll_ws_result.py +142 -0
- lockss/pyclient/poller/models/poller_detail.py +613 -0
- lockss/pyclient/poller/models/poller_pager.py +139 -0
- lockss/pyclient/poller/models/poller_summary.py +452 -0
- lockss/pyclient/poller/models/repair_data.py +176 -0
- lockss/pyclient/poller/models/repair_pager.py +139 -0
- lockss/pyclient/poller/models/repair_queue.py +249 -0
- lockss/pyclient/poller/models/repository_space_ws_result.py +113 -0
- lockss/pyclient/poller/models/repository_ws_result.py +113 -0
- lockss/pyclient/poller/models/tally_data.py +471 -0
- lockss/pyclient/poller/models/url_pager.py +139 -0
- lockss/pyclient/poller/models/vote_ws_result.py +142 -0
- lockss/pyclient/poller/models/voter_detail.py +701 -0
- lockss/pyclient/poller/models/voter_pager.py +139 -0
- lockss/pyclient/poller/models/voter_summary.py +284 -0
- lockss/pyclient/poller/rest.py +317 -0
- lockss/pyclient/rs/__init__.py +41 -0
- lockss/pyclient/rs/api/__init__.py +10 -0
- lockss/pyclient/rs/api/artifacts_api.py +988 -0
- lockss/pyclient/rs/api/aus_api.py +334 -0
- lockss/pyclient/rs/api/repo_api.py +379 -0
- lockss/pyclient/rs/api/status_api.py +120 -0
- lockss/pyclient/rs/api/wayback_api.py +386 -0
- lockss/pyclient/rs/api_client.py +632 -0
- lockss/pyclient/rs/configuration.py +247 -0
- lockss/pyclient/rs/models/__init__.py +31 -0
- lockss/pyclient/rs/models/api_status.py +344 -0
- lockss/pyclient/rs/models/archives_body.py +142 -0
- lockss/pyclient/rs/models/artifact.py +344 -0
- lockss/pyclient/rs/models/artifact_page_info.py +140 -0
- lockss/pyclient/rs/models/artifact_properties.py +344 -0
- lockss/pyclient/rs/models/artifacts_body.py +170 -0
- lockss/pyclient/rs/models/au_size.py +162 -0
- lockss/pyclient/rs/models/auid_page_info.py +140 -0
- lockss/pyclient/rs/models/error_result.py +164 -0
- lockss/pyclient/rs/models/import_status.py +298 -0
- lockss/pyclient/rs/models/page_info.py +229 -0
- lockss/pyclient/rs/models/repository_info.py +164 -0
- lockss/pyclient/rs/models/repository_statistics.py +112 -0
- lockss/pyclient/rs/models/storage_info.py +287 -0
- lockss/pyclient/rs/models/streaming_response_body.py +84 -0
- lockss/pyclient/rs/rest.py +317 -0
- lockss_pyclient-0.1.0.dev1.dist-info/LICENSE +27 -0
- lockss_pyclient-0.1.0.dev1.dist-info/METADATA +29 -0
- lockss_pyclient-0.1.0.dev1.dist-info/RECORD +148 -0
- lockss_pyclient-0.1.0.dev1.dist-info/WHEEL +4 -0
|
@@ -0,0 +1,45 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
# flake8: noqa
|
|
4
|
+
|
|
5
|
+
"""
|
|
6
|
+
LOCKSS Crawler Service REST API
|
|
7
|
+
|
|
8
|
+
REST API of the LOCKSS Crawler Service # noqa: E501
|
|
9
|
+
|
|
10
|
+
OpenAPI spec version: 2.0.0
|
|
11
|
+
Contact: lockss-support@lockss.org
|
|
12
|
+
Generated by: https://github.com/swagger-api/swagger-codegen.git
|
|
13
|
+
"""
|
|
14
|
+
|
|
15
|
+
from __future__ import absolute_import
|
|
16
|
+
|
|
17
|
+
# import apis into sdk package
|
|
18
|
+
from lockss.pyclient.crawler.api.crawlers_api import CrawlersApi
|
|
19
|
+
from lockss.pyclient.crawler.api.crawls_api import CrawlsApi
|
|
20
|
+
from lockss.pyclient.crawler.api.jobs_api import JobsApi
|
|
21
|
+
from lockss.pyclient.crawler.api.status_api import StatusApi
|
|
22
|
+
from lockss.pyclient.crawler.api.ws_api import WsApi
|
|
23
|
+
# import ApiClient
|
|
24
|
+
from lockss.pyclient.crawler.api_client import ApiClient
|
|
25
|
+
from lockss.pyclient.crawler.configuration import Configuration
|
|
26
|
+
# import models into sdk package
|
|
27
|
+
from lockss.pyclient.crawler.models.api_status import ApiStatus
|
|
28
|
+
from lockss.pyclient.crawler.models.counter import Counter
|
|
29
|
+
from lockss.pyclient.crawler.models.crawl_desc import CrawlDesc
|
|
30
|
+
from lockss.pyclient.crawler.models.crawl_job import CrawlJob
|
|
31
|
+
from lockss.pyclient.crawler.models.crawl_pager import CrawlPager
|
|
32
|
+
from lockss.pyclient.crawler.models.crawl_status import CrawlStatus
|
|
33
|
+
from lockss.pyclient.crawler.models.crawl_ws_result import CrawlWsResult
|
|
34
|
+
from lockss.pyclient.crawler.models.crawl_ws_result_pages_with_errors import CrawlWsResultPagesWithErrors
|
|
35
|
+
from lockss.pyclient.crawler.models.crawler_config import CrawlerConfig
|
|
36
|
+
from lockss.pyclient.crawler.models.crawler_status import CrawlerStatus
|
|
37
|
+
from lockss.pyclient.crawler.models.crawler_statuses import CrawlerStatuses
|
|
38
|
+
from lockss.pyclient.crawler.models.error_result import ErrorResult
|
|
39
|
+
from lockss.pyclient.crawler.models.job_pager import JobPager
|
|
40
|
+
from lockss.pyclient.crawler.models.job_status import JobStatus
|
|
41
|
+
from lockss.pyclient.crawler.models.mime_counter import MimeCounter
|
|
42
|
+
from lockss.pyclient.crawler.models.page_info import PageInfo
|
|
43
|
+
from lockss.pyclient.crawler.models.url_error import UrlError
|
|
44
|
+
from lockss.pyclient.crawler.models.url_info import UrlInfo
|
|
45
|
+
from lockss.pyclient.crawler.models.url_pager import UrlPager
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
from __future__ import absolute_import
|
|
2
|
+
|
|
3
|
+
# flake8: noqa
|
|
4
|
+
|
|
5
|
+
# import apis into api package
|
|
6
|
+
from lockss.pyclient.crawler.api.crawlers_api import CrawlersApi
|
|
7
|
+
from lockss.pyclient.crawler.api.crawls_api import CrawlsApi
|
|
8
|
+
from lockss.pyclient.crawler.api.jobs_api import JobsApi
|
|
9
|
+
from lockss.pyclient.crawler.api.status_api import StatusApi
|
|
10
|
+
from lockss.pyclient.crawler.api.ws_api import WsApi
|
|
@@ -0,0 +1,215 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
LOCKSS Crawler Service REST API
|
|
5
|
+
|
|
6
|
+
REST API of the LOCKSS Crawler Service # noqa: E501
|
|
7
|
+
|
|
8
|
+
OpenAPI spec version: 2.0.0
|
|
9
|
+
Contact: lockss-support@lockss.org
|
|
10
|
+
Generated by: https://github.com/swagger-api/swagger-codegen.git
|
|
11
|
+
"""
|
|
12
|
+
|
|
13
|
+
from __future__ import absolute_import
|
|
14
|
+
|
|
15
|
+
import re # noqa: F401
|
|
16
|
+
|
|
17
|
+
# python 2 and python 3 compatibility library
|
|
18
|
+
import six
|
|
19
|
+
|
|
20
|
+
from lockss.pyclient.crawler.api_client import ApiClient
|
|
21
|
+
|
|
22
|
+
|
|
23
|
+
class CrawlersApi(object):
|
|
24
|
+
"""NOTE: This class is auto generated by the swagger code generator program.
|
|
25
|
+
|
|
26
|
+
Do not edit the class manually.
|
|
27
|
+
Ref: https://github.com/swagger-api/swagger-codegen
|
|
28
|
+
"""
|
|
29
|
+
|
|
30
|
+
def __init__(self, api_client=None):
|
|
31
|
+
if api_client is None:
|
|
32
|
+
api_client = ApiClient()
|
|
33
|
+
self.api_client = api_client
|
|
34
|
+
|
|
35
|
+
def get_crawler_config(self, crawler_id, **kwargs): # noqa: E501
|
|
36
|
+
"""Return information about a crawler. # noqa: E501
|
|
37
|
+
|
|
38
|
+
Get information related to a installed crawler. # noqa: E501
|
|
39
|
+
This method makes a synchronous HTTP request by default. To make an
|
|
40
|
+
asynchronous HTTP request, please pass async_req=True
|
|
41
|
+
>>> thread = api.get_crawler_config(crawler_id, async_req=True)
|
|
42
|
+
>>> result = thread.get()
|
|
43
|
+
|
|
44
|
+
:param async_req bool
|
|
45
|
+
:param str crawler_id: Identifier for the crawler (required)
|
|
46
|
+
:return: CrawlerConfig
|
|
47
|
+
If the method is called asynchronously,
|
|
48
|
+
returns the request thread.
|
|
49
|
+
"""
|
|
50
|
+
kwargs['_return_http_data_only'] = True
|
|
51
|
+
if kwargs.get('async_req'):
|
|
52
|
+
return self.get_crawler_config_with_http_info(crawler_id, **kwargs) # noqa: E501
|
|
53
|
+
else:
|
|
54
|
+
(data) = self.get_crawler_config_with_http_info(crawler_id, **kwargs) # noqa: E501
|
|
55
|
+
return data
|
|
56
|
+
|
|
57
|
+
def get_crawler_config_with_http_info(self, crawler_id, **kwargs): # noqa: E501
|
|
58
|
+
"""Return information about a crawler. # noqa: E501
|
|
59
|
+
|
|
60
|
+
Get information related to a installed crawler. # noqa: E501
|
|
61
|
+
This method makes a synchronous HTTP request by default. To make an
|
|
62
|
+
asynchronous HTTP request, please pass async_req=True
|
|
63
|
+
>>> thread = api.get_crawler_config_with_http_info(crawler_id, async_req=True)
|
|
64
|
+
>>> result = thread.get()
|
|
65
|
+
|
|
66
|
+
:param async_req bool
|
|
67
|
+
:param str crawler_id: Identifier for the crawler (required)
|
|
68
|
+
:return: CrawlerConfig
|
|
69
|
+
If the method is called asynchronously,
|
|
70
|
+
returns the request thread.
|
|
71
|
+
"""
|
|
72
|
+
|
|
73
|
+
all_params = ['crawler_id'] # noqa: E501
|
|
74
|
+
all_params.append('async_req')
|
|
75
|
+
all_params.append('_return_http_data_only')
|
|
76
|
+
all_params.append('_preload_content')
|
|
77
|
+
all_params.append('_request_timeout')
|
|
78
|
+
|
|
79
|
+
params = locals()
|
|
80
|
+
for key, val in six.iteritems(params['kwargs']):
|
|
81
|
+
if key not in all_params:
|
|
82
|
+
raise TypeError(
|
|
83
|
+
"Got an unexpected keyword argument '%s'"
|
|
84
|
+
" to method get_crawler_config" % key
|
|
85
|
+
)
|
|
86
|
+
params[key] = val
|
|
87
|
+
del params['kwargs']
|
|
88
|
+
# verify the required parameter 'crawler_id' is set
|
|
89
|
+
if ('crawler_id' not in params or
|
|
90
|
+
params['crawler_id'] is None):
|
|
91
|
+
raise ValueError("Missing the required parameter `crawler_id` when calling `get_crawler_config`") # noqa: E501
|
|
92
|
+
|
|
93
|
+
collection_formats = {}
|
|
94
|
+
|
|
95
|
+
path_params = {}
|
|
96
|
+
if 'crawler_id' in params:
|
|
97
|
+
path_params['crawlerId'] = params['crawler_id'] # noqa: E501
|
|
98
|
+
|
|
99
|
+
query_params = []
|
|
100
|
+
|
|
101
|
+
header_params = {}
|
|
102
|
+
|
|
103
|
+
form_params = []
|
|
104
|
+
local_var_files = {}
|
|
105
|
+
|
|
106
|
+
body_params = None
|
|
107
|
+
# HTTP header `Accept`
|
|
108
|
+
header_params['Accept'] = self.api_client.select_header_accept(
|
|
109
|
+
['application/json']) # noqa: E501
|
|
110
|
+
|
|
111
|
+
# Authentication setting
|
|
112
|
+
auth_settings = ['basicAuth'] # noqa: E501
|
|
113
|
+
|
|
114
|
+
return self.api_client.call_api(
|
|
115
|
+
'/crawlers/{crawlerId}', 'GET',
|
|
116
|
+
path_params,
|
|
117
|
+
query_params,
|
|
118
|
+
header_params,
|
|
119
|
+
body=body_params,
|
|
120
|
+
post_params=form_params,
|
|
121
|
+
files=local_var_files,
|
|
122
|
+
response_type='CrawlerConfig', # noqa: E501
|
|
123
|
+
auth_settings=auth_settings,
|
|
124
|
+
async_req=params.get('async_req'),
|
|
125
|
+
_return_http_data_only=params.get('_return_http_data_only'),
|
|
126
|
+
_preload_content=params.get('_preload_content', True),
|
|
127
|
+
_request_timeout=params.get('_request_timeout'),
|
|
128
|
+
collection_formats=collection_formats)
|
|
129
|
+
|
|
130
|
+
def get_crawlers(self, **kwargs): # noqa: E501
|
|
131
|
+
"""Get the list of supported crawlers. # noqa: E501
|
|
132
|
+
|
|
133
|
+
Return the list of supported crawlers. # noqa: E501
|
|
134
|
+
This method makes a synchronous HTTP request by default. To make an
|
|
135
|
+
asynchronous HTTP request, please pass async_req=True
|
|
136
|
+
>>> thread = api.get_crawlers(async_req=True)
|
|
137
|
+
>>> result = thread.get()
|
|
138
|
+
|
|
139
|
+
:param async_req bool
|
|
140
|
+
:return: CrawlerStatuses
|
|
141
|
+
If the method is called asynchronously,
|
|
142
|
+
returns the request thread.
|
|
143
|
+
"""
|
|
144
|
+
kwargs['_return_http_data_only'] = True
|
|
145
|
+
if kwargs.get('async_req'):
|
|
146
|
+
return self.get_crawlers_with_http_info(**kwargs) # noqa: E501
|
|
147
|
+
else:
|
|
148
|
+
(data) = self.get_crawlers_with_http_info(**kwargs) # noqa: E501
|
|
149
|
+
return data
|
|
150
|
+
|
|
151
|
+
def get_crawlers_with_http_info(self, **kwargs): # noqa: E501
|
|
152
|
+
"""Get the list of supported crawlers. # noqa: E501
|
|
153
|
+
|
|
154
|
+
Return the list of supported crawlers. # noqa: E501
|
|
155
|
+
This method makes a synchronous HTTP request by default. To make an
|
|
156
|
+
asynchronous HTTP request, please pass async_req=True
|
|
157
|
+
>>> thread = api.get_crawlers_with_http_info(async_req=True)
|
|
158
|
+
>>> result = thread.get()
|
|
159
|
+
|
|
160
|
+
:param async_req bool
|
|
161
|
+
:return: CrawlerStatuses
|
|
162
|
+
If the method is called asynchronously,
|
|
163
|
+
returns the request thread.
|
|
164
|
+
"""
|
|
165
|
+
|
|
166
|
+
all_params = [] # noqa: E501
|
|
167
|
+
all_params.append('async_req')
|
|
168
|
+
all_params.append('_return_http_data_only')
|
|
169
|
+
all_params.append('_preload_content')
|
|
170
|
+
all_params.append('_request_timeout')
|
|
171
|
+
|
|
172
|
+
params = locals()
|
|
173
|
+
for key, val in six.iteritems(params['kwargs']):
|
|
174
|
+
if key not in all_params:
|
|
175
|
+
raise TypeError(
|
|
176
|
+
"Got an unexpected keyword argument '%s'"
|
|
177
|
+
" to method get_crawlers" % key
|
|
178
|
+
)
|
|
179
|
+
params[key] = val
|
|
180
|
+
del params['kwargs']
|
|
181
|
+
|
|
182
|
+
collection_formats = {}
|
|
183
|
+
|
|
184
|
+
path_params = {}
|
|
185
|
+
|
|
186
|
+
query_params = []
|
|
187
|
+
|
|
188
|
+
header_params = {}
|
|
189
|
+
|
|
190
|
+
form_params = []
|
|
191
|
+
local_var_files = {}
|
|
192
|
+
|
|
193
|
+
body_params = None
|
|
194
|
+
# HTTP header `Accept`
|
|
195
|
+
header_params['Accept'] = self.api_client.select_header_accept(
|
|
196
|
+
['application/json']) # noqa: E501
|
|
197
|
+
|
|
198
|
+
# Authentication setting
|
|
199
|
+
auth_settings = ['basicAuth'] # noqa: E501
|
|
200
|
+
|
|
201
|
+
return self.api_client.call_api(
|
|
202
|
+
'/crawlers', 'GET',
|
|
203
|
+
path_params,
|
|
204
|
+
query_params,
|
|
205
|
+
header_params,
|
|
206
|
+
body=body_params,
|
|
207
|
+
post_params=form_params,
|
|
208
|
+
files=local_var_files,
|
|
209
|
+
response_type='CrawlerStatuses', # noqa: E501
|
|
210
|
+
auth_settings=auth_settings,
|
|
211
|
+
async_req=params.get('async_req'),
|
|
212
|
+
_return_http_data_only=params.get('_return_http_data_only'),
|
|
213
|
+
_preload_content=params.get('_preload_content', True),
|
|
214
|
+
_request_timeout=params.get('_request_timeout'),
|
|
215
|
+
collection_formats=collection_formats)
|