scdataloader 1.2.1__tar.gz → 1.6.3__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {scdataloader-1.2.1 → scdataloader-1.6.3}/PKG-INFO +4 -3
- {scdataloader-1.2.1 → scdataloader-1.6.3}/pyproject.toml +13 -1
- scdataloader-1.6.3/scdataloader/VERSION +1 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/__main__.py +11 -4
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/utils.py +17 -10
- scdataloader-1.2.1/.cursorignore +0 -21
- scdataloader-1.2.1/.github/ISSUE_TEMPLATE/bug_report.md +0 -31
- scdataloader-1.2.1/.github/ISSUE_TEMPLATE/feature_request.md +0 -20
- scdataloader-1.2.1/.github/PULL_REQUEST_TEMPLATE.md +0 -15
- scdataloader-1.2.1/.github/dependabot.yml +0 -6
- scdataloader-1.2.1/.github/release_message.sh +0 -3
- scdataloader-1.2.1/.github/rename_project.sh +0 -36
- scdataloader-1.2.1/.github/workflows/main.yml +0 -43
- scdataloader-1.2.1/.github/workflows/release.yml +0 -48
- scdataloader-1.2.1/ABOUT_THIS_TEMPLATE.md +0 -70
- scdataloader-1.2.1/CONTRIBUTING.md +0 -113
- scdataloader-1.2.1/Containerfile +0 -5
- scdataloader-1.2.1/HISTORY.md +0 -296
- scdataloader-1.2.1/MANIFEST.in +0 -5
- scdataloader-1.2.1/Makefile +0 -95
- scdataloader-1.2.1/docs/collator.md +0 -4
- scdataloader-1.2.1/docs/datamodule.md +0 -4
- scdataloader-1.2.1/docs/dataset.md +0 -7
- scdataloader-1.2.1/docs/index.md +0 -182
- scdataloader-1.2.1/docs/notebooks/1_download_and_preprocess.ipynb +0 -901
- scdataloader-1.2.1/docs/notebooks/2_create_dataloader.ipynb +0 -341
- scdataloader-1.2.1/docs/preprocess.md +0 -13
- scdataloader-1.2.1/docs/scdataloader.drawio.png +0 -0
- scdataloader-1.2.1/docs/utils.md +0 -4
- scdataloader-1.2.1/mkdocs.yml +0 -36
- scdataloader-1.2.1/notebooks/additional.py +0 -24
- scdataloader-1.2.1/notebooks/finalize_data_loader.ipynb +0 -1611
- scdataloader-1.2.1/notebooks/finalize_data_loaderv2.ipynb +0 -2436
- scdataloader-1.2.1/notebooks/onto_rel.ipynb +0 -519
- scdataloader-1.2.1/notebooks/prepare_dataset.py +0 -96
- scdataloader-1.2.1/notebooks/rel_onto_tissues_age.ipynb +0 -494
- scdataloader-1.2.1/notebooks/reset_lamin.py +0 -26
- scdataloader-1.2.1/notebooks/speed_dataloader.ipynb +0 -191
- scdataloader-1.2.1/notebooks/work_on_dataloader_onto part 2.ipynb +0 -4629
- scdataloader-1.2.1/notebooks/work_on_dataloader_onto part 3.ipynb +0 -6106
- scdataloader-1.2.1/notebooks/work_on_dataloader_onto.ipynb +0 -3496
- scdataloader-1.2.1/poetry.lock +0 -5061
- scdataloader-1.2.1/scdataloader/VERSION +0 -1
- scdataloader-1.2.1/tests/__init__.py +0 -0
- scdataloader-1.2.1/tests/conftest.py +0 -26
- scdataloader-1.2.1/tests/test.h5ad +0 -0
- scdataloader-1.2.1/tests/test_base.py +0 -64
- scdataloader-1.2.1/uv.lock +0 -3157
- {scdataloader-1.2.1 → scdataloader-1.6.3}/.gitignore +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/LICENSE +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/README.md +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/__init__.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/base.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/collator.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/config.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/data.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/datamodule.py +0 -0
- {scdataloader-1.2.1 → scdataloader-1.6.3}/scdataloader/preprocess.py +0 -0
|
@@ -1,23 +1,24 @@
|
|
|
1
1
|
Metadata-Version: 2.3
|
|
2
2
|
Name: scdataloader
|
|
3
|
-
Version: 1.
|
|
3
|
+
Version: 1.6.3
|
|
4
4
|
Summary: a dataloader for single cell data in lamindb
|
|
5
5
|
Project-URL: repository, https://github.com/jkobject/scDataLoader
|
|
6
6
|
Author-email: jkobject <jkobject@gmail.com>
|
|
7
|
-
License
|
|
8
|
-
License-File: LICENSE
|
|
7
|
+
License: MIT
|
|
9
8
|
Keywords: dataloader,lamindb,pytorch,scPRINT,scRNAseq
|
|
10
9
|
Requires-Python: <3.11,>=3.10
|
|
11
10
|
Requires-Dist: anndata>=0.9.0
|
|
12
11
|
Requires-Dist: biomart>=0.9.0
|
|
13
12
|
Requires-Dist: cellxgene-census>=0.1.0
|
|
14
13
|
Requires-Dist: django>=4.0.0
|
|
14
|
+
Requires-Dist: harmonypy>=0.0.10
|
|
15
15
|
Requires-Dist: ipykernel>=6.20.0
|
|
16
16
|
Requires-Dist: lamindb[bionty]==0.76.12
|
|
17
17
|
Requires-Dist: leidenalg>=0.8.0
|
|
18
18
|
Requires-Dist: lightning>=2.0.0
|
|
19
19
|
Requires-Dist: matplotlib>=3.5.0
|
|
20
20
|
Requires-Dist: numpy>=1.26.0
|
|
21
|
+
Requires-Dist: palantir>=1.3.3
|
|
21
22
|
Requires-Dist: pandas>=2.0.0
|
|
22
23
|
Requires-Dist: scikit-misc>=0.5.0
|
|
23
24
|
Requires-Dist: seaborn>=0.11.0
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
[project]
|
|
2
2
|
name = "scdataloader"
|
|
3
|
-
version = "1.
|
|
3
|
+
version = "1.6.3"
|
|
4
4
|
description = "a dataloader for single cell data in lamindb"
|
|
5
5
|
authors = [
|
|
6
6
|
{name = "jkobject", email = "jkobject@gmail.com"}
|
|
@@ -25,6 +25,8 @@ dependencies = [
|
|
|
25
25
|
"leidenalg>=0.8.0",
|
|
26
26
|
"django>=4.0.0",
|
|
27
27
|
"scikit-misc>=0.5.0",
|
|
28
|
+
"palantir>=1.3.3",
|
|
29
|
+
"harmonypy>=0.0.10",
|
|
28
30
|
]
|
|
29
31
|
|
|
30
32
|
[project.optional-dependencies]
|
|
@@ -57,6 +59,16 @@ repository = "https://github.com/jkobject/scDataLoader"
|
|
|
57
59
|
requires = ["hatchling"]
|
|
58
60
|
build-backend = "hatchling.build"
|
|
59
61
|
|
|
62
|
+
[tool.hatch.build.targets.sdist]
|
|
63
|
+
only-include = [
|
|
64
|
+
"/scdataloader",
|
|
65
|
+
]
|
|
66
|
+
|
|
67
|
+
[tool.hatch.build.targets.wheel]
|
|
68
|
+
only-include = [
|
|
69
|
+
"/scdataloader",
|
|
70
|
+
]
|
|
71
|
+
|
|
60
72
|
[tool.hatch.metadata]
|
|
61
73
|
allow-direct-references = true
|
|
62
74
|
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
1.6.3
|
|
@@ -53,14 +53,14 @@ def main():
|
|
|
53
53
|
)
|
|
54
54
|
parser.add_argument(
|
|
55
55
|
"--filter_gene_by_counts",
|
|
56
|
-
type=
|
|
57
|
-
default=
|
|
56
|
+
type=int,
|
|
57
|
+
default=0,
|
|
58
58
|
help="Determines whether to filter genes by counts.",
|
|
59
59
|
)
|
|
60
60
|
parser.add_argument(
|
|
61
61
|
"--filter_cell_by_counts",
|
|
62
|
-
type=
|
|
63
|
-
default=
|
|
62
|
+
type=int,
|
|
63
|
+
default=0,
|
|
64
64
|
help="Determines whether to filter cells by counts.",
|
|
65
65
|
)
|
|
66
66
|
parser.add_argument(
|
|
@@ -153,6 +153,12 @@ def main():
|
|
|
153
153
|
default=False,
|
|
154
154
|
help="Determines whether to do postprocessing.",
|
|
155
155
|
)
|
|
156
|
+
parser.add_argument(
|
|
157
|
+
"--cache",
|
|
158
|
+
type=bool,
|
|
159
|
+
default=True,
|
|
160
|
+
help="Determines whether to cache the dataset.",
|
|
161
|
+
)
|
|
156
162
|
args = parser.parse_args()
|
|
157
163
|
|
|
158
164
|
# Load the collection
|
|
@@ -178,6 +184,7 @@ def main():
|
|
|
178
184
|
normalize_sum=args.normalize_sum,
|
|
179
185
|
subset_hvg=args.subset_hvg,
|
|
180
186
|
hvg_flavor=args.hvg_flavor,
|
|
187
|
+
cache=args.cache,
|
|
181
188
|
binning=args.binning,
|
|
182
189
|
result_binned_key=args.result_binned_key,
|
|
183
190
|
length_normalize=args.length_normalize,
|
|
@@ -127,15 +127,15 @@ def getBiomartTable(
|
|
|
127
127
|
|
|
128
128
|
cache_folder = os.path.expanduser(cache_folder)
|
|
129
129
|
createFoldersFor(cache_folder)
|
|
130
|
-
cachefile = os.path.join(cache_folder, ".biomart.
|
|
130
|
+
cachefile = os.path.join(cache_folder, ".biomart.parquet")
|
|
131
131
|
if useCache & os.path.isfile(cachefile):
|
|
132
132
|
print("fetching gene names from biomart cache")
|
|
133
|
-
res = pd.
|
|
133
|
+
res = pd.read_parquet(cachefile)
|
|
134
134
|
else:
|
|
135
135
|
print("downloading gene names from biomart")
|
|
136
136
|
|
|
137
137
|
res = _fetchFromServer(ensemble_server, attr + attributes, database=database)
|
|
138
|
-
res.
|
|
138
|
+
res.to_parquet(cachefile, index=False)
|
|
139
139
|
res.columns = attr + attributes
|
|
140
140
|
if type(res) is not type(pd.DataFrame()):
|
|
141
141
|
raise ValueError("should be a dataframe")
|
|
@@ -368,7 +368,14 @@ def load_genes(organisms: Union[str, list] = "NCBITaxon:9606"): # "NCBITaxon:10
|
|
|
368
368
|
genesdf["organism"] = organism
|
|
369
369
|
organismdf.append(genesdf)
|
|
370
370
|
organismdf = pd.concat(organismdf)
|
|
371
|
-
for col in [
|
|
371
|
+
for col in [
|
|
372
|
+
"source_id",
|
|
373
|
+
"run_id",
|
|
374
|
+
"created_by_id",
|
|
375
|
+
"updated_at",
|
|
376
|
+
"stable_id",
|
|
377
|
+
"created_at",
|
|
378
|
+
]:
|
|
372
379
|
if col in organismdf.columns:
|
|
373
380
|
organismdf.drop(columns=[col], inplace=True)
|
|
374
381
|
return organismdf
|
|
@@ -412,7 +419,7 @@ def populate_my_ontology(
|
|
|
412
419
|
# cell type
|
|
413
420
|
if celltypes is not None:
|
|
414
421
|
if len(celltypes) == 0:
|
|
415
|
-
bt.CellType.import_from_source()
|
|
422
|
+
bt.CellType.import_from_source(update=True)
|
|
416
423
|
else:
|
|
417
424
|
names = bt.CellType.public().df().index if not celltypes else celltypes
|
|
418
425
|
records = bt.CellType.from_values(names, field="ontology_id")
|
|
@@ -446,7 +453,7 @@ def populate_my_ontology(
|
|
|
446
453
|
# ethnicity
|
|
447
454
|
if ethnicities is not None:
|
|
448
455
|
if len(ethnicities) == 0:
|
|
449
|
-
bt.Ethnicity.import_from_source()
|
|
456
|
+
bt.Ethnicity.import_from_source(update=True)
|
|
450
457
|
else:
|
|
451
458
|
names = bt.Ethnicity.public().df().index if not ethnicities else ethnicities
|
|
452
459
|
records = bt.Ethnicity.from_values(names, field="ontology_id")
|
|
@@ -457,7 +464,7 @@ def populate_my_ontology(
|
|
|
457
464
|
# ExperimentalFactor
|
|
458
465
|
if assays is not None:
|
|
459
466
|
if len(assays) == 0:
|
|
460
|
-
bt.ExperimentalFactor.import_from_source()
|
|
467
|
+
bt.ExperimentalFactor.import_from_source(update=True)
|
|
461
468
|
else:
|
|
462
469
|
names = bt.ExperimentalFactor.public().df().index if not assays else assays
|
|
463
470
|
records = bt.ExperimentalFactor.from_values(names, field="ontology_id")
|
|
@@ -468,7 +475,7 @@ def populate_my_ontology(
|
|
|
468
475
|
# Tissue
|
|
469
476
|
if tissues is not None:
|
|
470
477
|
if len(tissues) == 0:
|
|
471
|
-
bt.Tissue.import_from_source()
|
|
478
|
+
bt.Tissue.import_from_source(update=True)
|
|
472
479
|
else:
|
|
473
480
|
names = bt.Tissue.public().df().index if not tissues else tissues
|
|
474
481
|
records = bt.Tissue.from_values(names, field="ontology_id")
|
|
@@ -477,7 +484,7 @@ def populate_my_ontology(
|
|
|
477
484
|
# DevelopmentalStage
|
|
478
485
|
if dev_stages is not None:
|
|
479
486
|
if len(dev_stages) == 0:
|
|
480
|
-
bt.DevelopmentalStage.import_from_source()
|
|
487
|
+
bt.DevelopmentalStage.import_from_source(update=True)
|
|
481
488
|
source = bt.PublicSource.filter(organism="mouse", name="mmusdv").last()
|
|
482
489
|
bt.DevelopmentalStage.import_from_source(source=source)
|
|
483
490
|
else:
|
|
@@ -493,7 +500,7 @@ def populate_my_ontology(
|
|
|
493
500
|
# Disease
|
|
494
501
|
if diseases is not None:
|
|
495
502
|
if len(diseases) == 0:
|
|
496
|
-
bt.Disease.import_from_source()
|
|
503
|
+
bt.Disease.import_from_source(update=True)
|
|
497
504
|
else:
|
|
498
505
|
names = bt.Disease.public().df().index if not diseases else diseases
|
|
499
506
|
records = bt.Disease.from_values(names, field="ontology_id")
|
scdataloader-1.2.1/.cursorignore
DELETED
|
@@ -1,31 +0,0 @@
|
|
|
1
|
-
---
|
|
2
|
-
name: Bug report
|
|
3
|
-
about: Create a report to help us improve
|
|
4
|
-
title: ''
|
|
5
|
-
labels: bug, help wanted
|
|
6
|
-
assignees: ''
|
|
7
|
-
|
|
8
|
-
---
|
|
9
|
-
|
|
10
|
-
**Describe the bug**
|
|
11
|
-
A clear and concise description of what the bug is.
|
|
12
|
-
|
|
13
|
-
**To Reproduce**
|
|
14
|
-
Steps to reproduce the behavior:
|
|
15
|
-
1. Go to '...'
|
|
16
|
-
2. Click on '....'
|
|
17
|
-
3. Scroll down to '....'
|
|
18
|
-
4. See error
|
|
19
|
-
|
|
20
|
-
**Expected behavior**
|
|
21
|
-
A clear and concise description of what you expected to happen.
|
|
22
|
-
|
|
23
|
-
**Screenshots**
|
|
24
|
-
If applicable, add screenshots to help explain your problem.
|
|
25
|
-
|
|
26
|
-
**Desktop (please complete the following information):**
|
|
27
|
-
- OS: [e.g. iOS]
|
|
28
|
-
- Version [e.g. 22]
|
|
29
|
-
|
|
30
|
-
**Additional context**
|
|
31
|
-
Add any other context about the problem here.
|
|
@@ -1,20 +0,0 @@
|
|
|
1
|
-
---
|
|
2
|
-
name: Feature request
|
|
3
|
-
about: Suggest an idea for this project
|
|
4
|
-
title: ''
|
|
5
|
-
labels: enhancement, question
|
|
6
|
-
assignees: ''
|
|
7
|
-
|
|
8
|
-
---
|
|
9
|
-
|
|
10
|
-
**Is your feature request related to a problem? Please describe.**
|
|
11
|
-
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
|
|
12
|
-
|
|
13
|
-
**Describe the solution you'd like**
|
|
14
|
-
A clear and concise description of what you want to happen.
|
|
15
|
-
|
|
16
|
-
**Describe alternatives you've considered**
|
|
17
|
-
A clear and concise description of any alternative solutions or features you've considered.
|
|
18
|
-
|
|
19
|
-
**Additional context**
|
|
20
|
-
Add any other context or screenshots about the feature request here.
|
|
@@ -1,15 +0,0 @@
|
|
|
1
|
-
### Summary :memo:
|
|
2
|
-
_Write an overview about it._
|
|
3
|
-
|
|
4
|
-
### Details
|
|
5
|
-
_Describe more what you did on changes._
|
|
6
|
-
1. (...)
|
|
7
|
-
2. (...)
|
|
8
|
-
|
|
9
|
-
### Bugfixes :bug: (delete if dind't have any)
|
|
10
|
-
-
|
|
11
|
-
|
|
12
|
-
### Checks
|
|
13
|
-
- [ ] Closed #798
|
|
14
|
-
- [ ] Tested Changes
|
|
15
|
-
- [ ] Stakeholder Approval
|
|
@@ -1,36 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/env bash
|
|
2
|
-
while getopts a:n:u:d: flag
|
|
3
|
-
do
|
|
4
|
-
case "${flag}" in
|
|
5
|
-
a) author=${OPTARG};;
|
|
6
|
-
n) name=${OPTARG};;
|
|
7
|
-
u) urlname=${OPTARG};;
|
|
8
|
-
d) description=${OPTARG};;
|
|
9
|
-
esac
|
|
10
|
-
done
|
|
11
|
-
|
|
12
|
-
echo "Author: $author";
|
|
13
|
-
echo "Project Name: $name";
|
|
14
|
-
echo "Project URL name: $urlname";
|
|
15
|
-
echo "Description: $description";
|
|
16
|
-
|
|
17
|
-
echo "Renaming project..."
|
|
18
|
-
|
|
19
|
-
original_author="jkobject"
|
|
20
|
-
original_name="scdataloader"
|
|
21
|
-
original_urlname="scDataLoader"
|
|
22
|
-
original_description="Awesome scdataloader created by jkobject"
|
|
23
|
-
# for filename in $(find . -name "*.*")
|
|
24
|
-
for filename in $(git ls-files)
|
|
25
|
-
do
|
|
26
|
-
sed -i "s/$original_author/$author/g" $filename
|
|
27
|
-
sed -i "s/$original_name/$name/g" $filename
|
|
28
|
-
sed -i "s/$original_urlname/$urlname/g" $filename
|
|
29
|
-
sed -i "s/$original_description/$description/g" $filename
|
|
30
|
-
echo "Renamed $filename"
|
|
31
|
-
done
|
|
32
|
-
|
|
33
|
-
mv scdataloader $name
|
|
34
|
-
|
|
35
|
-
# This command runs only once on GHA!
|
|
36
|
-
rm -rf .github/template.yml
|
|
@@ -1,43 +0,0 @@
|
|
|
1
|
-
# This is a basic workflow to help you get started with Actions
|
|
2
|
-
|
|
3
|
-
name: CI
|
|
4
|
-
|
|
5
|
-
# Controls when the workflow will run
|
|
6
|
-
on:
|
|
7
|
-
# Triggers the workflow on push or pull request events but only for the main branch
|
|
8
|
-
push:
|
|
9
|
-
branches: [main]
|
|
10
|
-
pull_request:
|
|
11
|
-
branches: [main]
|
|
12
|
-
|
|
13
|
-
# Allows you to run this workflow manually from the Actions tab
|
|
14
|
-
workflow_dispatch:
|
|
15
|
-
|
|
16
|
-
jobs:
|
|
17
|
-
ci:
|
|
18
|
-
strategy:
|
|
19
|
-
fail-fast: true
|
|
20
|
-
matrix:
|
|
21
|
-
python-version: ["3.10"]
|
|
22
|
-
os: [ubuntu-latest]
|
|
23
|
-
runs-on: ${{ matrix.os }}
|
|
24
|
-
steps:
|
|
25
|
-
- uses: actions/checkout@v4
|
|
26
|
-
- name: Install uv
|
|
27
|
-
uses: astral-sh/setup-uv@v3
|
|
28
|
-
- uses: actions/setup-python@v5
|
|
29
|
-
with:
|
|
30
|
-
python-version: ${{ matrix.python-version }}
|
|
31
|
-
- name: Install project
|
|
32
|
-
run: make virtualenv
|
|
33
|
-
continue-on-error: false
|
|
34
|
-
- name: Run linter
|
|
35
|
-
run: make lint
|
|
36
|
-
continue-on-error: false
|
|
37
|
-
- name: Run tests
|
|
38
|
-
run: make test
|
|
39
|
-
- name: "Upload coverage to Codecov"
|
|
40
|
-
uses: codecov/codecov-action@v4
|
|
41
|
-
with:
|
|
42
|
-
fail_ci_if_error: true
|
|
43
|
-
token: ${{ secrets.CODECOV_TOKEN }}
|
|
@@ -1,48 +0,0 @@
|
|
|
1
|
-
name: Upload Python Package
|
|
2
|
-
|
|
3
|
-
on:
|
|
4
|
-
push:
|
|
5
|
-
# Sequence of patterns matched against refs/tags
|
|
6
|
-
tags:
|
|
7
|
-
- "*" # Push events to matching v*, i.e. v1.0, v20.15.10
|
|
8
|
-
|
|
9
|
-
# Allows you to run this workflow manually from the Actions tab
|
|
10
|
-
workflow_dispatch:
|
|
11
|
-
|
|
12
|
-
jobs:
|
|
13
|
-
release:
|
|
14
|
-
name: Create Release
|
|
15
|
-
runs-on: ubuntu-latest
|
|
16
|
-
permissions:
|
|
17
|
-
contents: write
|
|
18
|
-
steps:
|
|
19
|
-
- uses: actions/checkout@v4
|
|
20
|
-
with:
|
|
21
|
-
# by default, it uses a depth of 1
|
|
22
|
-
# this fetches all history so that we can read each commit
|
|
23
|
-
fetch-depth: 0
|
|
24
|
-
- name: Generate Changelog
|
|
25
|
-
run: .github/release_message.sh > release_message.md
|
|
26
|
-
- name: Release
|
|
27
|
-
uses: softprops/action-gh-release@v1
|
|
28
|
-
with:
|
|
29
|
-
body_path: release_message.md
|
|
30
|
-
|
|
31
|
-
deploy:
|
|
32
|
-
needs: release
|
|
33
|
-
strategy:
|
|
34
|
-
fail-fast: true
|
|
35
|
-
matrix:
|
|
36
|
-
python-version: ["3.10"]
|
|
37
|
-
os: [ubuntu-latest]
|
|
38
|
-
runs-on: ${{ matrix.os }}
|
|
39
|
-
steps:
|
|
40
|
-
- uses: actions/checkout@v4
|
|
41
|
-
- name: Install uv
|
|
42
|
-
uses: astral-sh/setup-uv@v3
|
|
43
|
-
- uses: actions/setup-python@v5
|
|
44
|
-
with:
|
|
45
|
-
python-version: ${{ matrix.python-version }}
|
|
46
|
-
- name: Build and publish
|
|
47
|
-
run: |
|
|
48
|
-
uv build && uv publish -u __token__ -p ${{ secrets.POETRY_TOKEN }}
|
|
@@ -1,70 +0,0 @@
|
|
|
1
|
-
## Structure
|
|
2
|
-
|
|
3
|
-
Lets take a look at the structure of this template:
|
|
4
|
-
|
|
5
|
-
```text
|
|
6
|
-
├── Containerfile # The file to build a container using buildah or docker
|
|
7
|
-
├── CONTRIBUTING.md # Onboarding instructions for new contributors
|
|
8
|
-
├── docs # Documentation site (add more .md files here)
|
|
9
|
-
│ └── index.md # The index page for the docs site
|
|
10
|
-
├── .github # Github metadata for repository
|
|
11
|
-
│ ├── release_message.sh # A script to generate a release message
|
|
12
|
-
│ └── workflows # The CI pipeline for Github Actions
|
|
13
|
-
├── .gitignore # A list of files to ignore when pushing to Github
|
|
14
|
-
├── HISTORY.md # Auto generated list of changes to the project
|
|
15
|
-
├── LICENSE # The license for the project
|
|
16
|
-
├── Makefile # A collection of utilities to manage the project
|
|
17
|
-
├── MANIFEST.in # A list of files to include in a package
|
|
18
|
-
├── mkdocs.yml # Configuration for documentation site
|
|
19
|
-
├── scdataloader # The main python package for the project
|
|
20
|
-
│ ├── base.py # The base module for the project
|
|
21
|
-
│ ├── __init__.py # This tells Python that this is a package
|
|
22
|
-
│ ├── __main__.py # The entry point for the project
|
|
23
|
-
│ └── VERSION # The version for the project is kept in a static file
|
|
24
|
-
├── README.md # The main readme for the project
|
|
25
|
-
├── setup.py # The setup.py file for installing and packaging the project
|
|
26
|
-
├── requirements.txt # An empty file to hold the requirements for the project
|
|
27
|
-
├── requirements-test.txt # List of requirements for testing and devlopment
|
|
28
|
-
├── setup.py # The setup.py file for installing and packaging the project
|
|
29
|
-
└── tests # Unit tests for the project (add mote tests files here)
|
|
30
|
-
├── conftest.py # Configuration, hooks and fixtures for pytest
|
|
31
|
-
├── __init__.py # This tells Python that this is a test package
|
|
32
|
-
└── test_base.py # The base test case for the project
|
|
33
|
-
```
|
|
34
|
-
|
|
35
|
-
### Why to include `tests`, `history` and `Containerfile` as part of the release?
|
|
36
|
-
|
|
37
|
-
The `MANIFEST.in` file is used to include the files in the release, once the
|
|
38
|
-
project is released to PyPI all the files listed on MANIFEST.in will be included
|
|
39
|
-
even if the files are static or not related to Python.
|
|
40
|
-
|
|
41
|
-
Some build systems such as RPM, DEB, AUR for some Linux distributions, and also
|
|
42
|
-
internal repackaging systems tends to run the tests before the packaging is performed.
|
|
43
|
-
|
|
44
|
-
The Containerfile can be useful to provide a safer execution environment for
|
|
45
|
-
the project when running on a testing environment.
|
|
46
|
-
|
|
47
|
-
I added those files to make it easier for packaging in different formats.
|
|
48
|
-
|
|
49
|
-
## The Makefile
|
|
50
|
-
|
|
51
|
-
All the utilities for the template and project are on the Makefile
|
|
52
|
-
|
|
53
|
-
```bash
|
|
54
|
-
❯ make
|
|
55
|
-
Usage: make <target>
|
|
56
|
-
|
|
57
|
-
Targets:
|
|
58
|
-
help: ## Show the help.
|
|
59
|
-
install: ## Install the project in dev mode.
|
|
60
|
-
fmt: ## Format code using black & isort.
|
|
61
|
-
lint: ## Run pep8, black, mypy linters.
|
|
62
|
-
test: lint ## Run tests and generate coverage report.
|
|
63
|
-
watch: ## Run tests on every change.
|
|
64
|
-
clean: ## Clean unused files.
|
|
65
|
-
virtualenv: ## Create a virtual environment.
|
|
66
|
-
release: ## Create a new tag for release.
|
|
67
|
-
docs: ## Build the documentation.
|
|
68
|
-
switch-to-poetry: ## Switch to poetry package manager.
|
|
69
|
-
init: ## Initialize the project based on an application template.
|
|
70
|
-
```
|
|
@@ -1,113 +0,0 @@
|
|
|
1
|
-
# How to develop on this project
|
|
2
|
-
|
|
3
|
-
scdataloader welcomes contributions from the community.
|
|
4
|
-
|
|
5
|
-
**You need PYTHON3!**
|
|
6
|
-
|
|
7
|
-
This instructions are for linux base systems. (Linux, MacOS, BSD, etc.)
|
|
8
|
-
## Setting up your own fork of this repo.
|
|
9
|
-
|
|
10
|
-
- On github interface click on `Fork` button.
|
|
11
|
-
- Clone your fork of this repo. `git clone git@github.com:YOUR_GIT_USERNAME/scDataLoader.git`
|
|
12
|
-
- Enter the directory `cd scDataLoader`
|
|
13
|
-
- Add upstream repo `git remote add upstream https://github.com/jkobject/scDataLoader`
|
|
14
|
-
|
|
15
|
-
## Setting up your own virtual environment
|
|
16
|
-
|
|
17
|
-
Run `make virtualenv` to create a virtual environment.
|
|
18
|
-
then activate it with `source .venv/bin/activate`.
|
|
19
|
-
|
|
20
|
-
## Install the project in develop mode
|
|
21
|
-
|
|
22
|
-
Run `make install` to install the project in develop mode.
|
|
23
|
-
|
|
24
|
-
## Run the tests to ensure everything is working
|
|
25
|
-
|
|
26
|
-
Run `make test` to run the tests.
|
|
27
|
-
|
|
28
|
-
## Create a new branch to work on your contribution
|
|
29
|
-
|
|
30
|
-
Run `git checkout -b my_contribution`
|
|
31
|
-
|
|
32
|
-
## Make your changes
|
|
33
|
-
|
|
34
|
-
Edit the files using your preferred editor. (we recommend VIM or VSCode)
|
|
35
|
-
|
|
36
|
-
## Format the code
|
|
37
|
-
|
|
38
|
-
Run `make fmt` to format the code.
|
|
39
|
-
|
|
40
|
-
## Run the linter
|
|
41
|
-
|
|
42
|
-
Run `make lint` to run the linter.
|
|
43
|
-
|
|
44
|
-
## Test your changes
|
|
45
|
-
|
|
46
|
-
Run `make test` to run the tests.
|
|
47
|
-
|
|
48
|
-
Ensure code coverage report shows `100%` coverage, add tests to your PR.
|
|
49
|
-
|
|
50
|
-
## Build the docs locally
|
|
51
|
-
|
|
52
|
-
Run `make docs` to build the docs.
|
|
53
|
-
|
|
54
|
-
Ensure your new changes are documented.
|
|
55
|
-
|
|
56
|
-
## Commit your changes
|
|
57
|
-
|
|
58
|
-
This project uses [conventional git commit messages](https://www.conventionalcommits.org/en/v1.0.0/).
|
|
59
|
-
|
|
60
|
-
Example: `fix(package): update setup.py arguments 🎉` (emojis are fine too)
|
|
61
|
-
|
|
62
|
-
## Push your changes to your fork
|
|
63
|
-
|
|
64
|
-
Run `git push origin my_contribution`
|
|
65
|
-
|
|
66
|
-
## Submit a pull request
|
|
67
|
-
|
|
68
|
-
On github interface, click on `Pull Request` button.
|
|
69
|
-
|
|
70
|
-
Wait CI to run and one of the developers will review your PR.
|
|
71
|
-
## Makefile utilities
|
|
72
|
-
|
|
73
|
-
This project comes with a `Makefile` that contains a number of useful utility.
|
|
74
|
-
|
|
75
|
-
```bash
|
|
76
|
-
❯ make
|
|
77
|
-
Usage: make <target>
|
|
78
|
-
|
|
79
|
-
Targets:
|
|
80
|
-
help: ## Show the help.
|
|
81
|
-
install: ## Install the project in dev mode.
|
|
82
|
-
fmt: ## Format code using black & isort.
|
|
83
|
-
lint: ## Run pep8, black, mypy linters.
|
|
84
|
-
test: lint ## Run tests and generate coverage report.
|
|
85
|
-
watch: ## Run tests on every change.
|
|
86
|
-
clean: ## Clean unused files.
|
|
87
|
-
virtualenv: ## Create a virtual environment.
|
|
88
|
-
release: ## Create a new tag for release.
|
|
89
|
-
docs: ## Build the documentation.
|
|
90
|
-
switch-to-poetry: ## Switch to poetry package manager.
|
|
91
|
-
init: ## Initialize the project based on an application template.
|
|
92
|
-
```
|
|
93
|
-
|
|
94
|
-
## Making a new release
|
|
95
|
-
|
|
96
|
-
This project uses [semantic versioning](https://semver.org/) and tags releases with `X.Y.Z`
|
|
97
|
-
Every time a new tag is created and pushed to the remote repo, github actions will
|
|
98
|
-
automatically create a new release on github and trigger a release on PyPI.
|
|
99
|
-
|
|
100
|
-
For this to work you need to setup a secret called `PIPY_API_TOKEN` on the project settings>secrets,
|
|
101
|
-
this token can be generated on [pypi.org](https://pypi.org/account/).
|
|
102
|
-
|
|
103
|
-
To trigger a new release all you need to do is.
|
|
104
|
-
|
|
105
|
-
1. If you have changes to add to the repo
|
|
106
|
-
* Make your changes following the steps described above.
|
|
107
|
-
* Commit your changes following the [conventional git commit messages](https://www.conventionalcommits.org/en/v1.0.0/).
|
|
108
|
-
2. Run the tests to ensure everything is working.
|
|
109
|
-
4. Run `make release` to create a new tag and push it to the remote repo.
|
|
110
|
-
|
|
111
|
-
the `make release` will ask you the version number to create the tag, ex: type `0.1.1` when you are asked.
|
|
112
|
-
|
|
113
|
-
> **CAUTION**: The make release will change local changelog files and commit all the unstaged changes you have.
|