Skip to content

Commit 470bc4a

Browse files
committed
using .connect
1 parent fea5d5e commit 470bc4a

7 files changed

Lines changed: 18 additions & 26 deletions

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -211,7 +211,7 @@ from scdataloader.preprocess import LaminPreprocessor, additional_postprocess, a
211211
# preprocess datasets
212212
DESCRIPTION='preprocessed by scDataLoader'
213213

214-
cx_dataset = ln.Collection.using(instance="laminlabs/cellxgene").filter(name="cellxgene-census", version='2023-12-15').one()
214+
cx_dataset = ln.Collection.connect(instance="laminlabs/cellxgene").filter(name="cellxgene-census", version='2023-12-15').one()
215215
cx_dataset, len(cx_dataset.artifacts.all())
216216

217217
# (OPTIONAL) if you want to do you preprocessing on a slurm cluster without internet connections,

docs/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -201,7 +201,7 @@ from scdataloader.preprocess import LaminPreprocessor, additional_postprocess, a
201201
# preprocess datasets
202202
DESCRIPTION='preprocessed by scDataLoader'
203203

204-
cx_dataset = ln.Collection.using(instance="laminlabs/cellxgene").filter(name="cellxgene-census", version='2023-12-15').one()
204+
cx_dataset = ln.Collection.connect(instance="laminlabs/cellxgene").filter(name="cellxgene-census", version='2023-12-15').one()
205205
cx_dataset, len(cx_dataset.artifacts.all())
206206

207207
# (OPTIONAL) if you want to do you preprocessing on a slurm cluster without internet connections,

docs/notebooks/1_download_and_preprocess.ipynb

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,7 @@
9999
},
100100
{
101101
"cell_type": "code",
102-
"execution_count": 5,
102+
"execution_count": null,
103103
"metadata": {},
104104
"outputs": [
105105
{
@@ -116,12 +116,12 @@
116116
}
117117
],
118118
"source": [
119-
"list(ln.Collection.using(instance=\"laminlabs/cellxgene\").filter(name=\"cellxgene-census\").all())"
119+
"list(ln.Collection.connect(instance=\"laminlabs/cellxgene\").filter(name=\"cellxgene-census\").all())"
120120
]
121121
},
122122
{
123123
"cell_type": "code",
124-
"execution_count": 7,
124+
"execution_count": null,
125125
"metadata": {},
126126
"outputs": [
127127
{
@@ -144,7 +144,7 @@
144144
}
145145
],
146146
"source": [
147-
"cx_dataset = ln.Collection.using(instance=\"laminlabs/cellxgene\").filter(name=\"cellxgene-census\", version='2023-12-15').one()\n",
147+
"cx_dataset = ln.Collection.connect(instance=\"laminlabs/cellxgene\").filter(name=\"cellxgene-census\", version='2023-12-15').one()\n",
148148
"cx_dataset, len(cx_dataset.artifacts.all())"
149149
]
150150
},

docs/notebooks/2_create_dataloader.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -112,7 +112,7 @@
112112
"\n",
113113
"example:\n",
114114
"```python\n",
115-
"dataset = ln.Collection.using(\"laminlabs/cellxgene\").one()\n",
115+
"dataset = ln.Collection.connect(\"laminlabs/cellxgene\").one()\n",
116116
"```"
117117
]
118118
},

notebooks/update_lamin_or_cellxgene.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1677,7 +1677,7 @@
16771677
"import lamindb as ln\n",
16781678
"\n",
16791679
"files = (\n",
1680-
" ln.Collection.using(instance=\"laminlabs/cellxgene\")\n",
1680+
" ln.Collection.connect(instance=\"laminlabs/cellxgene\")\n",
16811681
" .filter(name=\"cellxgene-census\")\n",
16821682
" .first()\n",
16831683
" .artifacts.all()\n",
@@ -1720,7 +1720,7 @@
17201720
"outputs": [],
17211721
"source": [
17221722
"cx_dataset = (\n",
1723-
" ln.Collection.using(instance=\"laminlabs/cellxgene\")\n",
1723+
" ln.Collection.connect(instance=\"laminlabs/cellxgene\")\n",
17241724
" .filter(name=\"cellxgene-census\", version=\"2024-11-10\")\n",
17251725
" .one()\n",
17261726
")"
@@ -1753,7 +1753,7 @@
17531753
],
17541754
"source": [
17551755
"cx_dataset_prev = (\n",
1756-
" ln.Collection.using(instance=\"laminlabs/cellxgene\")\n",
1756+
" ln.Collection.connect(instance=\"laminlabs/cellxgene\")\n",
17571757
" .filter(name=\"cellxgene-census\", version=\"2023-12-15\")\n",
17581758
" .one()\n",
17591759
")\n",

scdataloader/__main__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -228,7 +228,7 @@ def main():
228228
# return
229229
if args.instance is not None:
230230
collection = (
231-
ln.Collection.using(instance=args.instance)
231+
ln.Collection.connect(instance=args.instance)
232232
.filter(key=args.name, version=args.version)
233233
.first()
234234
)

scdataloader/utils.py

Lines changed: 7 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -454,19 +454,19 @@ def _adding_scbasecamp_genes(
454454
):
455455
if len(species) == 0:
456456
species = set(
457-
bt.Organism.using("laminlabs/arc-virtual-cell-atlas")
457+
bt.Organism.connect("laminlabs/arc-virtual-cell-atlas")
458458
.to_dataframe(limit=None)
459459
.ontology_id
460460
) - set(["NCBITaxon:10090", "NCBITaxon:9606"])
461461
species = list(species)
462462
for i in set(
463-
bt.Organism.using("laminlabs/arc-virtual-cell-atlas")
463+
bt.Organism.connect("laminlabs/arc-virtual-cell-atlas")
464464
.to_dataframe(limit=None)
465465
.ontology_id
466466
) - set(bt.Organism.filter().to_dataframe(limit=None).ontology_id):
467467
print(i)
468468
rec = (
469-
bt.Organism.using("laminlabs/arc-virtual-cell-atlas")
469+
bt.Organism.connect("laminlabs/arc-virtual-cell-atlas")
470470
.filter(ontology_id=i)
471471
.first()
472472
)
@@ -487,7 +487,7 @@ def _adding_scbasecamp_genes(
487487
for i in species:
488488
print(i)
489489
df = (
490-
bt.Gene.using("laminlabs/arc-virtual-cell-atlas")
490+
bt.Gene.connect("laminlabs/arc-virtual-cell-atlas")
491491
.filter(organism__ontology_id=i)
492492
.all()
493493
.to_dataframe(limit=None)
@@ -633,11 +633,7 @@ def populate_my_ontology(
633633
if len(tissues) == 0:
634634
bt.Tissue.import_source()
635635
else:
636-
names = (
637-
bt.Tissue.public().to_dataframe().index
638-
if not tissues
639-
else tissues
640-
)
636+
names = bt.Tissue.public().to_dataframe().index if not tissues else tissues
641637
records = bt.Tissue.from_values(names, field="ontology_id")
642638
ln.save(records)
643639
elem = bt.Tissue(name="unknown", ontology_id="unknown").save()
@@ -663,9 +659,7 @@ def populate_my_ontology(
663659
bt.Disease.import_source()
664660
else:
665661
names = (
666-
bt.Disease.public().to_dataframe().index
667-
if not diseases
668-
else diseases
662+
bt.Disease.public().to_dataframe().index if not diseases else diseases
669663
)
670664
records = bt.Disease.from_values(names, field="ontology_id")
671665
ln.save(records)
@@ -675,9 +669,7 @@ def populate_my_ontology(
675669
for organism in genes_from:
676670
# convert onto to name
677671
organism = bt.Organism.filter(ontology_id=organism).one().name
678-
names = bt.Gene.public(organism=organism).to_dataframe()[
679-
"ensembl_gene_id"
680-
]
672+
names = bt.Gene.public(organism=organism).to_dataframe()["ensembl_gene_id"]
681673

682674
# Process names in blocks of 10,000 elements
683675
block_size = 10000

0 commit comments

Comments
 (0)