From 146c3d6d62797c7c3843ef9e948fbc5d1692a5e5 Mon Sep 17 00:00:00 2001 From: Sourcery AI <> Date: Sat, 21 Jan 2023 05:06:09 +0000 Subject: [PATCH] 'Refactored by Sourcery' --- scraper/main.py | 7 +++---- scraper/tool.py | 10 +++------- 2 files changed, 6 insertions(+), 11 deletions(-) diff --git a/scraper/main.py b/scraper/main.py index c32247c..c3c4326 100644 --- a/scraper/main.py +++ b/scraper/main.py @@ -24,10 +24,9 @@ def run(): temp = fetch(exclude) if not temp: break - else: - result += temp - exclude += [x["objectID"] for x in temp] - count += 1 + result += temp + exclude += [x["objectID"] for x in temp] + count += 1 logger.info(f"Got {len(result)} new data.") diff --git a/scraper/tool.py b/scraper/tool.py index 5369948..ebe4317 100644 --- a/scraper/tool.py +++ b/scraper/tool.py @@ -14,14 +14,12 @@ def fetch(filter_object_id=[]): "requests": [ { "indexName": "job_postings", - "params": f'query=&hitsPerPage=1000&facetFilters=[["city.work_country_name:Indonesia"]]', + "params": 'query=&hitsPerPage=1000&facetFilters=[["city.work_country_name:Indonesia"]]', } ] } - if not filter_object_id: - pass - else: + if filter_object_id: # Construct filter to exclude passed objectID filter = f"&filters=NOT objectID:{filter_object_id[0]}" @@ -31,9 +29,7 @@ def fetch(filter_object_id=[]): data["requests"][0]["params"] += filter req = requests.post(os.environ["REQUEST_URL"], data=json.dumps(data)) - res = req.json()["results"][0]["hits"] - - return res + return req.json()["results"][0]["hits"] def get_job_ids():