1919)
2020def test_compute_merge_unit_groups (sorting_analyzer_with_splits , preset ):
2121
22- job_kwargs = dict (n_jobs = - 1 )
2322 sorting_analyzer , num_unit_splitted , other_ids = sorting_analyzer_with_splits
2423
2524 if preset is not None :
@@ -39,7 +38,6 @@ def test_compute_merge_unit_groups(sorting_analyzer_with_splits, preset):
3938 # adaptative_window_thresh=0.5,
4039 # firing_contamination_balance=1.5,
4140 extra_outputs = True ,
42- ** job_kwargs ,
4341 )
4442 if preset == "x_contaminations" :
4543 assert len (merge_unit_groups ) == num_unit_splitted
@@ -54,51 +52,43 @@ def test_compute_merge_unit_groups(sorting_analyzer_with_splits, preset):
5452 sorting_analyzer ,
5553 preset = preset ,
5654 steps = ["num_spikes" , "snr" , "remove_contaminated" , "unit_locations" ],
57- ** job_kwargs ,
5855 )
5956
6057
6158@pytest .mark .parametrize (
6259 "preset" , ["x_contaminations" , "feature_neighbors" , "temporal_splits" , "similarity_correlograms" , "slay" ]
6360)
6461def test_compute_merge_unit_groups_multi_segment (sorting_analyzer_multi_segment_for_curation , preset ):
65- job_kwargs = dict (n_jobs = - 1 )
6662 sorting_analyzer = sorting_analyzer_multi_segment_for_curation
6763 print (sorting_analyzer )
6864
6965 merge_unit_groups = compute_merge_unit_groups (
7066 sorting_analyzer ,
7167 preset = preset ,
72- ** job_kwargs ,
7368 )
7469
7570
7671def test_auto_merge_units (sorting_analyzer_for_curation ):
7772 recording = sorting_analyzer_for_curation .recording
78- job_kwargs = dict (n_jobs = - 1 )
7973 new_sorting , _ = split_sorting_by_times (sorting_analyzer_for_curation )
8074 new_sorting_analyzer = create_sorting_analyzer (new_sorting , recording , format = "memory" )
81- merged_analyzer = auto_merge_units (new_sorting_analyzer , presets = "x_contaminations" , ** job_kwargs )
75+ merged_analyzer = auto_merge_units (new_sorting_analyzer , presets = "x_contaminations" )
8276 assert len (merged_analyzer .unit_ids ) < len (new_sorting_analyzer .unit_ids )
8377
8478 step_merged_analyzer = auto_merge_units (
8579 new_sorting_analyzer ,
8680 presets = None ,
8781 steps = ["num_spikes" , "remove_contaminated" , "unit_locations" , "template_similarity" , "quality_score" ],
8882 steps_params = {"num_spikes" : {"min_spikes" : 150 }},
89- ** job_kwargs ,
9083 )
9184 assert len (step_merged_analyzer .unit_ids ) < len (new_sorting_analyzer .unit_ids )
9285
9386
9487def test_auto_merge_units_iterative (sorting_analyzer_for_curation ):
9588 recording = sorting_analyzer_for_curation .recording
96- job_kwargs = dict (n_jobs = - 1 )
9789 new_sorting , _ = split_sorting_by_times (sorting_analyzer_for_curation )
9890 new_sorting_analyzer = create_sorting_analyzer (new_sorting , recording , format = "memory" )
99- merged_analyzer = auto_merge_units (
100- new_sorting_analyzer , presets = ["x_contaminations" , "x_contaminations" ], ** job_kwargs
101- )
91+ merged_analyzer = auto_merge_units (new_sorting_analyzer , presets = ["x_contaminations" , "x_contaminations" ])
10292 assert len (merged_analyzer .unit_ids ) < len (new_sorting_analyzer .unit_ids )
10393
10494
0 commit comments