data-forager 0.1.2__tar.gz → 0.1.3__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {data_forager-0.1.2 → data_forager-0.1.3}/PKG-INFO +1 -1
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/indexers/tokenization_indexer.py +98 -29
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager.egg-info/PKG-INFO +1 -1
- {data_forager-0.1.2 → data_forager-0.1.3}/pyproject.toml +1 -1
- {data_forager-0.1.2 → data_forager-0.1.3}/LICENSE +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/README.md +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/__init__.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/datasets/__init__.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/datasets/common.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/datasets/jsonl.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/datasets/tokens.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/index_stores/__init__.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/index_stores/common.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/index_stores/fs_based.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/indexers/__init__.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/indexers/jsonl_indexer.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/indexers/text_lines.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/sample_index.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/unzip_files.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager/utils.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager.egg-info/SOURCES.txt +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager.egg-info/dependency_links.txt +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager.egg-info/requires.txt +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/data_forager.egg-info/top_level.txt +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/setup.cfg +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/tests/test_jsonl_indexing.py +0 -0
- {data_forager-0.1.2 → data_forager-0.1.3}/tests/test_tokenizing_indexing_jsonl.py +0 -0
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
from typing import Callable, List, Optional
|
|
2
2
|
|
|
3
|
+
import logging
|
|
3
4
|
import os
|
|
4
5
|
from pathlib import Path
|
|
5
6
|
|
|
@@ -8,7 +9,11 @@ import json
|
|
|
8
9
|
import numpy as np
|
|
9
10
|
|
|
10
11
|
from basics.base import Base
|
|
12
|
+
from basics.logging import get_logger
|
|
11
13
|
|
|
14
|
+
module_logger = get_logger(os.path.basename(__file__))
|
|
15
|
+
|
|
16
|
+
from data_forager.index_stores.common import IndexStoreInterface
|
|
12
17
|
from data_forager.index_stores.fs_based import IndexStore as FSBasedIndexStore
|
|
13
18
|
from data_forager.indexers.text_lines import SampleData, FileTextLinesIndexer, SampleGeneratorInterface
|
|
14
19
|
from data_forager.utils import find_files_recursive, natural_sort
|
|
@@ -24,61 +29,125 @@ def get_text_from_jsonl(jsonl_bytes: bytes, text_key: str = "text", text_encodin
|
|
|
24
29
|
|
|
25
30
|
|
|
26
31
|
def create_tokenize_and_index_jsonl_text_func(
|
|
27
|
-
input_base_path: str,
|
|
28
32
|
tokenizer_func: TokenizerFunc,
|
|
29
33
|
eos_idx: int,
|
|
34
|
+
input_base_path: Optional[str] = None,
|
|
35
|
+
input_file_paths: Optional[List[str]] = None,
|
|
36
|
+
output_base_path: Optional[str] = None,
|
|
37
|
+
index_store: Optional[IndexStoreInterface] = None,
|
|
30
38
|
process_text_line_func: Optional[ProcessTextLineFunc] = None,
|
|
39
|
+
logger: Optional[logging.Logger] = None,
|
|
31
40
|
name: Optional[str] = None,
|
|
32
41
|
**sample_generator_kwargs,
|
|
33
42
|
) -> FileTextLinesIndexer:
|
|
34
43
|
"""
|
|
35
|
-
Create
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
*
|
|
44
|
+
Create a pipeline to tokenize text from JSONL files and create an index for random access.
|
|
45
|
+
|
|
46
|
+
The pipeline:
|
|
47
|
+
* Tokenizes text from input JSONL objects
|
|
48
|
+
* Stores the token data in bin files under "tokenized-samples" folder
|
|
49
|
+
* Stores index data under "index" folder
|
|
39
50
|
|
|
40
51
|
Usage:
|
|
41
|
-
|
|
52
|
+
```python
|
|
42
53
|
import tiktoken
|
|
43
54
|
|
|
44
55
|
enc = tiktoken.get_encoding("gpt2")
|
|
45
56
|
def tokenize_text(text: str) -> List[int]:
|
|
46
|
-
return
|
|
57
|
+
return enc.encode_ordinary(text)
|
|
58
|
+
|
|
59
|
+
# Option 1: Scan directory for JSONL files, output to same directory
|
|
60
|
+
indexer = create_tokenize_and_index_jsonl_text_func(
|
|
61
|
+
tokenizer_func=tokenize_text,
|
|
62
|
+
eos_idx=enc.eot_token,
|
|
63
|
+
input_base_path='./data',
|
|
64
|
+
sample_size=1024,
|
|
65
|
+
)
|
|
47
66
|
|
|
48
|
-
|
|
49
|
-
|
|
67
|
+
# Option 2: Explicit input files and output path
|
|
68
|
+
indexer = create_tokenize_and_index_jsonl_text_func(
|
|
50
69
|
tokenizer_func=tokenize_text,
|
|
51
|
-
|
|
70
|
+
eos_idx=enc.eot_token,
|
|
71
|
+
input_file_paths=['./data/train.jsonl'],
|
|
72
|
+
output_base_path='./output',
|
|
73
|
+
sample_size=1024,
|
|
52
74
|
)
|
|
53
75
|
|
|
54
|
-
#
|
|
55
|
-
|
|
76
|
+
# Run tokenization and indexing
|
|
77
|
+
indexer()
|
|
78
|
+
```
|
|
56
79
|
|
|
57
|
-
:param input_base_path: Path to directory containing JSONL files (searched recursively).
|
|
58
80
|
:param tokenizer_func: Function used to tokenize text.
|
|
59
|
-
:param eos_idx: EOS token index, known by the used Tokenizer
|
|
81
|
+
:param eos_idx: EOS token index, known by the used Tokenizer.
|
|
82
|
+
:param input_base_path: Path to directory containing JSONL files (searched recursively).
|
|
83
|
+
Used as fallback for output if `output_base_path` is not provided.
|
|
84
|
+
:param input_file_paths: List of file paths to process. If provided, these are used
|
|
85
|
+
instead of scanning `input_base_path` for JSONL files.
|
|
86
|
+
:param output_base_path: Base path for output (index and tokenized samples).
|
|
87
|
+
If not provided, `input_base_path` is used.
|
|
88
|
+
:param index_store: Index store to use. If provided, this is used instead of
|
|
89
|
+
creating a new FSBasedIndexStore.
|
|
60
90
|
:param process_text_line_func: Function used to process text lines.
|
|
61
91
|
By default, this converts input JSON lines to dicts and returns the "text" field.
|
|
62
92
|
See function get_text_from_jsonl().
|
|
63
|
-
:param
|
|
64
|
-
:param name:
|
|
93
|
+
:param logger: Logger to use. If not provided, uses module logger.
|
|
94
|
+
:param name: Name of the indexer, used for logging purposes.
|
|
95
|
+
:param sample_generator_kwargs: Other kwargs passed to TokenizedSampleGenerator
|
|
96
|
+
(e.g., sample_size, token_dtype, base_output_path).
|
|
97
|
+
|
|
98
|
+
:raises ValueError: If both `input_base_path` and `input_file_paths` are None.
|
|
99
|
+
:raises ValueError: If `index_store` is None and both `output_base_path` and
|
|
100
|
+
`input_base_path` are None.
|
|
65
101
|
|
|
66
|
-
:return: FileTextLinesIndexer instance that can be
|
|
67
|
-
|
|
102
|
+
:return: FileTextLinesIndexer instance that can be called to run tokenization
|
|
103
|
+
and indexing.
|
|
68
104
|
"""
|
|
69
|
-
if
|
|
70
|
-
|
|
105
|
+
if logger is None:
|
|
106
|
+
logger = module_logger
|
|
71
107
|
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
108
|
+
# Validate input source
|
|
109
|
+
if input_base_path is None and input_file_paths is None:
|
|
110
|
+
raise ValueError(
|
|
111
|
+
"Either input_base_path or input_file_paths must be provided"
|
|
112
|
+
)
|
|
113
|
+
|
|
114
|
+
# Determine output base path
|
|
115
|
+
effective_output_base_path = output_base_path or input_base_path
|
|
79
116
|
|
|
80
|
-
#
|
|
81
|
-
|
|
117
|
+
# Validate output destination
|
|
118
|
+
if index_store is None and effective_output_base_path is None:
|
|
119
|
+
raise ValueError(
|
|
120
|
+
"Either index_store, output_base_path, or input_base_path must be provided "
|
|
121
|
+
"to determine where to store the index"
|
|
122
|
+
)
|
|
123
|
+
|
|
124
|
+
logger.info(f"Output base path: {effective_output_base_path}")
|
|
125
|
+
|
|
126
|
+
if process_text_line_func is None:
|
|
127
|
+
process_text_line_func = get_text_from_jsonl
|
|
128
|
+
|
|
129
|
+
if index_store is None:
|
|
130
|
+
index_store = FSBasedIndexStore(
|
|
131
|
+
base_path=effective_output_base_path,
|
|
132
|
+
)
|
|
133
|
+
|
|
134
|
+
if input_file_paths is None:
|
|
135
|
+
logger.info(f"Scanning for JSONL files in: {input_base_path}")
|
|
136
|
+
input_file_paths = find_files_recursive(
|
|
137
|
+
input_base_path,
|
|
138
|
+
extension_patterns=['*.jsonl', '*.JSONL']
|
|
139
|
+
)
|
|
140
|
+
# Assuming numbered files
|
|
141
|
+
input_file_paths = natural_sort(input_file_paths)
|
|
142
|
+
logger.info(f"Found {len(input_file_paths)} JSONL file(s)")
|
|
143
|
+
|
|
144
|
+
# Set default base_output_path for tokenized samples if not provided in kwargs
|
|
145
|
+
if 'base_output_path' not in sample_generator_kwargs:
|
|
146
|
+
default_base_output_path = os.path.join(
|
|
147
|
+
effective_output_base_path, "tokenized-samples"
|
|
148
|
+
)
|
|
149
|
+
logger.info(f"Tokenized samples output path: {default_base_output_path}")
|
|
150
|
+
sample_generator_kwargs['base_output_path'] = default_base_output_path
|
|
82
151
|
|
|
83
152
|
sample_generator = TokenizedSampleGenerator(
|
|
84
153
|
process_text_line_func=process_text_line_func,
|
|
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|
|
4
4
|
|
|
5
5
|
[project]
|
|
6
6
|
name = "data-forager"
|
|
7
|
-
version = "0.1.
|
|
7
|
+
version = "0.1.3"
|
|
8
8
|
description = "Enabling random access to large datasets on disk for PyTorch training and other use cases"
|
|
9
9
|
readme = "README.md"
|
|
10
10
|
license = "MIT"
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|