Skip to content

Commit 37ff412

Browse files
authored
Remove n_jobs=-1 from curation tests (#4449)
1 parent bb7c7ad commit 37ff412

3 files changed

Lines changed: 9 additions & 21 deletions

File tree

src/spikeinterface/curation/tests/common.py

Lines changed: 6 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -5,8 +5,6 @@
55
from spikeinterface.curation import train_model
66
from pathlib import Path
77

8-
job_kwargs = dict(n_jobs=-1)
9-
108
extensions = [
119
"noise_levels",
1210
"random_spikes",
@@ -21,7 +19,6 @@
2119

2220

2321
def make_sorting_analyzer(sparse=True, num_units=5, durations=[300.0]):
24-
job_kwargs = dict(n_jobs=-1)
2522
recording, sorting = generate_ground_truth_recording(
2623
durations=durations,
2724
sampling_frequency=30000.0,
@@ -38,15 +35,17 @@ def make_sorting_analyzer(sparse=True, num_units=5, durations=[300.0]):
3835
sorting = sorting.rename_units(new_unit_ids=unit_ids_as_integers)
3936

4037
sorting_analyzer = create_sorting_analyzer(
41-
sorting=sorting, recording=recording, format="memory", sparse=sparse, **job_kwargs
38+
sorting=sorting,
39+
recording=recording,
40+
format="memory",
41+
sparse=sparse,
4242
)
43-
sorting_analyzer.compute(extensions, **job_kwargs)
43+
sorting_analyzer.compute(extensions)
4444

4545
return sorting_analyzer
4646

4747

4848
def make_sorting_analyzer_with_splits(sorting_analyzer, num_unit_splitted=1, num_split=2):
49-
job_kwargs = dict(n_jobs=-1)
5049
sorting = sorting_analyzer.sorting
5150

5251
split_ids = sorting.unit_ids[:num_unit_splitted]
@@ -61,7 +60,7 @@ def make_sorting_analyzer_with_splits(sorting_analyzer, num_unit_splitted=1, num
6160
sorting_analyzer_with_splits = create_sorting_analyzer(
6261
sorting=sorting_with_split, recording=sorting_analyzer.recording, format="memory", sparse=True
6362
)
64-
sorting_analyzer_with_splits.compute(extensions, **job_kwargs)
63+
sorting_analyzer_with_splits.compute(extensions)
6564

6665
return sorting_analyzer_with_splits, num_unit_splitted, other_ids
6766

src/spikeinterface/curation/tests/test_auto_merge.py

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@
1919
)
2020
def test_compute_merge_unit_groups(sorting_analyzer_with_splits, preset):
2121

22-
job_kwargs = dict(n_jobs=-1)
2322
sorting_analyzer, num_unit_splitted, other_ids = sorting_analyzer_with_splits
2423

2524
if preset is not None:
@@ -39,7 +38,6 @@ def test_compute_merge_unit_groups(sorting_analyzer_with_splits, preset):
3938
# adaptative_window_thresh=0.5,
4039
# firing_contamination_balance=1.5,
4140
extra_outputs=True,
42-
**job_kwargs,
4341
)
4442
if preset == "x_contaminations":
4543
assert len(merge_unit_groups) == num_unit_splitted
@@ -54,51 +52,43 @@ def test_compute_merge_unit_groups(sorting_analyzer_with_splits, preset):
5452
sorting_analyzer,
5553
preset=preset,
5654
steps=["num_spikes", "snr", "remove_contaminated", "unit_locations"],
57-
**job_kwargs,
5855
)
5956

6057

6158
@pytest.mark.parametrize(
6259
"preset", ["x_contaminations", "feature_neighbors", "temporal_splits", "similarity_correlograms", "slay"]
6360
)
6461
def test_compute_merge_unit_groups_multi_segment(sorting_analyzer_multi_segment_for_curation, preset):
65-
job_kwargs = dict(n_jobs=-1)
6662
sorting_analyzer = sorting_analyzer_multi_segment_for_curation
6763
print(sorting_analyzer)
6864

6965
merge_unit_groups = compute_merge_unit_groups(
7066
sorting_analyzer,
7167
preset=preset,
72-
**job_kwargs,
7368
)
7469

7570

7671
def test_auto_merge_units(sorting_analyzer_for_curation):
7772
recording = sorting_analyzer_for_curation.recording
78-
job_kwargs = dict(n_jobs=-1)
7973
new_sorting, _ = split_sorting_by_times(sorting_analyzer_for_curation)
8074
new_sorting_analyzer = create_sorting_analyzer(new_sorting, recording, format="memory")
81-
merged_analyzer = auto_merge_units(new_sorting_analyzer, presets="x_contaminations", **job_kwargs)
75+
merged_analyzer = auto_merge_units(new_sorting_analyzer, presets="x_contaminations")
8276
assert len(merged_analyzer.unit_ids) < len(new_sorting_analyzer.unit_ids)
8377

8478
step_merged_analyzer = auto_merge_units(
8579
new_sorting_analyzer,
8680
presets=None,
8781
steps=["num_spikes", "remove_contaminated", "unit_locations", "template_similarity", "quality_score"],
8882
steps_params={"num_spikes": {"min_spikes": 150}},
89-
**job_kwargs,
9083
)
9184
assert len(step_merged_analyzer.unit_ids) < len(new_sorting_analyzer.unit_ids)
9285

9386

9487
def test_auto_merge_units_iterative(sorting_analyzer_for_curation):
9588
recording = sorting_analyzer_for_curation.recording
96-
job_kwargs = dict(n_jobs=-1)
9789
new_sorting, _ = split_sorting_by_times(sorting_analyzer_for_curation)
9890
new_sorting_analyzer = create_sorting_analyzer(new_sorting, recording, format="memory")
99-
merged_analyzer = auto_merge_units(
100-
new_sorting_analyzer, presets=["x_contaminations", "x_contaminations"], **job_kwargs
101-
)
91+
merged_analyzer = auto_merge_units(new_sorting_analyzer, presets=["x_contaminations", "x_contaminations"])
10292
assert len(merged_analyzer.unit_ids) < len(new_sorting_analyzer.unit_ids)
10393

10494

src/spikeinterface/curation/tests/test_remove_redundant.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,9 @@ def test_remove_redundant_units(sorting_analyzer_for_curation):
1616
# print(sorting.unit_ids)
1717
# print(sorting_with_dup.unit_ids)
1818

19-
job_kwargs = dict(n_jobs=-1)
2019
sorting_analyzer = create_sorting_analyzer(sorting_with_dup, recording, format="memory")
2120
sorting_analyzer.compute("random_spikes")
22-
sorting_analyzer.compute("waveforms", **job_kwargs)
21+
sorting_analyzer.compute("waveforms")
2322
sorting_analyzer.compute("templates")
2423

2524
for remove_strategy in ("max_spikes", "minimum_shift", "highest_amplitude"):

0 commit comments

Comments
 (0)