Skip to content

Commit f8c1473

Browse files
committed
possibly the last of the matches
Signed-off-by: John Seekins <john@robot-house.us>
1 parent e8c287d commit f8c1473

2 files changed

Lines changed: 23 additions & 1 deletion

File tree

ice_scrapers/utils.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -166,6 +166,7 @@ def repair_name(name: str, locality: str) -> Tuple[str, bool]:
166166
"replace": "WASHINGTON COUNTY JAIL",
167167
"locality": "HURRICANE",
168168
},
169+
{"match": "ETOWAH COUNTY JAIL (ALABAMA)", "replace": "ETOWAH COUNTY JAIL", "locality": "GADSDEN"},
169170
]
170171
cleaned = False
171172
for m in matches:

ice_scrapers/vera_data.py

Lines changed: 22 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -157,6 +157,27 @@ def _vera_name_fixes(name: str, city: str) -> Tuple[str, bool]:
157157
"replace": "Chittenden Regional Correctional Facility",
158158
"city": "South Burlington",
159159
},
160+
{
161+
"match": "NW Regional Corrections Center",
162+
"replace": "Northwest Regional Corrections Center",
163+
"city": "Crookston",
164+
},
165+
{
166+
"match": "Lasalle ICE Processing Center",
167+
"replace": "Central Louisiana ICE Processing Center (CLIPC)",
168+
"city": "Jena",
169+
},
170+
{
171+
"match": "La Salle Co Regional Det. Center",
172+
"replace": "La Salle County Regional Detention Center",
173+
"city": "Encinal",
174+
},
175+
{
176+
"match": "Hancock Co Pub Sfty Cplx",
177+
"replace": "Hancock County Public Safety Complex",
178+
"city": "Bay St. Louis",
179+
},
180+
{"match": "Brooks County Jail (Contract)", "replace": "Brooks County Jail", "city": "Falfurrias"},
160181
]
161182
fixed = False
162183
for m in matches:
@@ -196,7 +217,7 @@ def collect_vera_facility_data(facilities_data: dict, keep_sheet: bool = True, f
196217
df = polars.read_csv(has_header=True, raise_if_empty=True, source=filename, use_pyarrow=True)
197218
if df.is_empty():
198219
raise ValueError("Empty CSV loaded somehow! %s", df)
199-
# first step to removing duplicates is easy
220+
# first step to removing duplicates is easy, but unlikely to actually filter anything
200221
df = df.unique()
201222
logger.debug("Extracted data: %s", df)
202223
"""

0 commit comments

Comments
 (0)