datamule 2.1.1__tar.gz → 2.1.3__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {datamule-2.1.1 → datamule-2.1.3}/PKG-INFO +1 -1
- {datamule-2.1.1 → datamule-2.1.3}/datamule/datamule/downloader.py +5 -1
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/document.py +31 -10
- {datamule-2.1.1 → datamule-2.1.3}/datamule/mapping_dicts/html_mapping_dicts.py +2 -2
- {datamule-2.1.1 → datamule-2.1.3}/datamule/submission.py +6 -6
- {datamule-2.1.1 → datamule-2.1.3}/datamule.egg-info/PKG-INFO +1 -1
- {datamule-2.1.1 → datamule-2.1.3}/setup.py +1 -1
- {datamule-2.1.1 → datamule-2.1.3}/datamule/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/config.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/data/listed_filer_metadata.csv +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/datamule/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/datamule/datamule_lookup.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/datamule/datamule_mysql_rds.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/datamule/sec_connector.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_13fhr.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_25nse.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_informationtable.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_npx.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_ownership.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_proxyvotingrecord.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_sbsef.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/tables_sdr.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/document/tables/utils.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/helper.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/index.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/mapping_dicts/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/mapping_dicts/txt_mapping_dicts.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/mapping_dicts/xml_mapping_dicts.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/package_updater.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/portfolio.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/portfolio_compression_utils.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/infrastructure/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/infrastructure/submissions_metadata.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/downloader.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/eftsquery.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/monitor.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/streamer.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/submissions/textsearch.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/utils.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/xbrl/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/xbrl/downloadcompanyfacts.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/xbrl/filter_xbrl.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/xbrl/streamcompanyfacts.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sec/xbrl/xbrlmonitor.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/seclibrary/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/seclibrary/bq.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/sheet.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/utils/__init__.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/utils/construct_submissions_data.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule/utils/format_accession.py +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule.egg-info/SOURCES.txt +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule.egg-info/dependency_links.txt +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule.egg-info/requires.txt +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/datamule.egg-info/top_level.txt +0 -0
- {datamule-2.1.1 → datamule-2.1.3}/setup.cfg +0 -0
@@ -23,7 +23,11 @@ from ..utils.format_accession import format_accession
|
|
23
23
|
# could be cleaned up
|
24
24
|
|
25
25
|
# Set up logging
|
26
|
-
logging.basicConfig(
|
26
|
+
logging.basicConfig(
|
27
|
+
level=logging.INFO,
|
28
|
+
format='%(asctime)s - %(levelname)s - %(message)s',
|
29
|
+
handlers=logging.getLogger().handlers,
|
30
|
+
)
|
27
31
|
logger = logging.getLogger(__name__)
|
28
32
|
|
29
33
|
|
@@ -12,6 +12,7 @@ from selectolax.parser import HTMLParser
|
|
12
12
|
from pathlib import Path
|
13
13
|
import webbrowser
|
14
14
|
from secsgml.utils import bytes_to_str
|
15
|
+
import tempfile
|
15
16
|
|
16
17
|
from .tables.tables import Tables
|
17
18
|
|
@@ -36,18 +37,19 @@ class Document:
|
|
36
37
|
# this will be filled by parsed
|
37
38
|
self._data = None
|
38
39
|
self._tables = None
|
40
|
+
self._text = None
|
39
41
|
|
40
42
|
|
41
43
|
|
42
44
|
#_load_text_content
|
43
45
|
def _preprocess_txt_content(self):
|
44
|
-
|
46
|
+
self._text = self.content.decode().translate(str.maketrans({
|
45
47
|
'\xa0': ' ', '\u2003': ' ',
|
46
48
|
'\u2018': "'", '\u2019': "'",
|
47
49
|
'\u201c': '"', '\u201d': '"'
|
48
50
|
}))
|
49
51
|
|
50
|
-
#
|
52
|
+
# needs work
|
51
53
|
def _preprocess_html_content(self):
|
52
54
|
parser = HTMLParser(self.content,detect_encoding=True,decode_errors='ignore')
|
53
55
|
|
@@ -95,7 +97,7 @@ class Document:
|
|
95
97
|
while '\n\n\n' in text:
|
96
98
|
text = text.replace('\n\n\n', '\n\n')
|
97
99
|
|
98
|
-
|
100
|
+
self._text = text.translate(str.maketrans({
|
99
101
|
'\xa0': ' ', '\u2003': ' ',
|
100
102
|
'\u2018': "'", '\u2019': "'",
|
101
103
|
'\u201c': '"', '\u201d': '"'
|
@@ -116,7 +118,7 @@ class Document:
|
|
116
118
|
mapping_dict = None
|
117
119
|
|
118
120
|
if self.extension == '.txt':
|
119
|
-
content = self.
|
121
|
+
content = self.text
|
120
122
|
if self.type == '10-Q':
|
121
123
|
mapping_dict = dict_10q
|
122
124
|
elif self.type == '10-K':
|
@@ -224,6 +226,15 @@ class Document:
|
|
224
226
|
self.parse()
|
225
227
|
return self._data
|
226
228
|
|
229
|
+
@property
|
230
|
+
def text(self):
|
231
|
+
if self._text is None:
|
232
|
+
if self.extension in ['.htm','.html']:
|
233
|
+
self._preprocess_html_content()
|
234
|
+
elif self.extension == '.txt':
|
235
|
+
self._preprocess_txt_content()
|
236
|
+
return self._text
|
237
|
+
|
227
238
|
def write_json(self, output_filename=None):
|
228
239
|
if not self.data:
|
229
240
|
self.parse()
|
@@ -308,18 +319,28 @@ class Document:
|
|
308
319
|
self.parse()
|
309
320
|
|
310
321
|
if not self.data:
|
311
|
-
|
312
|
-
webbrowser.open('file://' + str(self.path))
|
313
|
-
else:
|
314
|
-
pass
|
322
|
+
pass
|
315
323
|
else:
|
316
324
|
visualize_dict(self.data)
|
317
325
|
|
318
|
-
|
326
|
+
# alpha feature
|
327
|
+
def open(self):
|
328
|
+
"""Open the document. Experimental. Creates copy in temp, rather than use tar path for now."""
|
329
|
+
if self.extension in ['.htm', '.html','.txt','.jpg','.png', '.pdf']:
|
330
|
+
# Create a temporary file with the content and open it
|
331
|
+
|
332
|
+
with tempfile.NamedTemporaryFile(mode='wb', suffix=self.extension, delete=False) as f:
|
333
|
+
f.write(self.content)
|
334
|
+
temp_path = f.name
|
335
|
+
webbrowser.open('file://' + temp_path)
|
336
|
+
else:
|
337
|
+
print(f"Cannot open files with extension {self.extension}")
|
338
|
+
|
339
|
+
def get_section(self, title=None, title_regex=None,title_class=None, format='dict'):
|
319
340
|
if not self.data:
|
320
341
|
self.parse()
|
321
342
|
|
322
|
-
result = get_title(self.data,title)
|
343
|
+
result = get_title(self.data,title=title,title_regex=title_regex,title_class=title_class)
|
323
344
|
|
324
345
|
if format == 'text':
|
325
346
|
result = [item[1] for item in result]
|
@@ -1,7 +1,7 @@
|
|
1
1
|
dict_10k_html = {
|
2
2
|
('part',r'^part\s*([ivx]+)$') : 0,
|
3
3
|
('signatures',r'^signatures?\.*$') : 0,
|
4
|
-
('item',r'^item\s*(\d+)\.?([a-z])?') : 1,
|
4
|
+
('item',r'^item\s*(\d+)\.?([a-z])?(?![a-z])') : 1,
|
5
5
|
}
|
6
6
|
dict_10q_html = dict_10k_html
|
7
7
|
|
@@ -48,7 +48,7 @@ dict_10d_html = dict_10k_html
|
|
48
48
|
|
49
49
|
dict_20f_html = {
|
50
50
|
('part',r'^part\s*([ivx]+)') : 0,
|
51
|
-
('item',r'^item\s*(\d+)\.?([a-z])?') : 1,
|
51
|
+
('item',r'^item\s*(\d+)\.?([a-z])?(?![a-z])') : 1,
|
52
52
|
('letter',r'\d*\.?([a-z])') : 2,
|
53
53
|
('signatures',r'^signatures?\.*$') : 0,
|
54
54
|
}
|
@@ -163,8 +163,8 @@ class Submission:
|
|
163
163
|
content = zstd.ZstdDecompressor().decompress(content)
|
164
164
|
|
165
165
|
# Decode text files
|
166
|
-
if extension in ['.htm', '.html', '.txt', '.xml']:
|
167
|
-
|
166
|
+
# if extension in ['.htm', '.html', '.txt', '.xml']:
|
167
|
+
# content = content.decode('utf-8', errors='replace')
|
168
168
|
|
169
169
|
document_path = f"{self.batch_tar_path}::{self.accession_prefix}/{filename}"
|
170
170
|
|
@@ -197,8 +197,8 @@ class Submission:
|
|
197
197
|
content = zstd.ZstdDecompressor().decompress(content)
|
198
198
|
|
199
199
|
# Decode text files
|
200
|
-
if extension in ['.htm', '.html', '.txt', '.xml']:
|
201
|
-
|
200
|
+
# if extension in ['.htm', '.html', '.txt', '.xml']:
|
201
|
+
# content = content.decode('utf-8', errors='replace')
|
202
202
|
|
203
203
|
document_path = f"{self.path}::{actual_filename}"
|
204
204
|
|
@@ -219,8 +219,8 @@ class Submission:
|
|
219
219
|
content = zstd.ZstdDecompressor().decompress(content)
|
220
220
|
|
221
221
|
# Decode text files
|
222
|
-
if extension in ['.htm', '.html', '.txt', '.xml']:
|
223
|
-
|
222
|
+
# if extension in ['.htm', '.html', '.txt', '.xml']:
|
223
|
+
# content = content.decode('utf-8', errors='replace')
|
224
224
|
|
225
225
|
return Document(
|
226
226
|
type=doc['type'],
|
@@ -32,7 +32,7 @@ if not os.path.exists(file_path):
|
|
32
32
|
setup(
|
33
33
|
name="datamule",
|
34
34
|
author="John Friedman",
|
35
|
-
version="2.1.
|
35
|
+
version="2.1.3",
|
36
36
|
description="Work with SEC submissions at scale.",
|
37
37
|
packages=find_packages(include=['datamule', 'datamule.*']),
|
38
38
|
url="https://github.com/john-friedman/datamule-python",
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|