From e76d29807d9c9aa788c239b23afc47d0888d1269 Mon Sep 17 00:00:00 2001 From: cynthia Date: Sat, 5 Apr 2025 12:15:49 +0000 Subject: [PATCH] make the verified csv file optional argument --- scripts/crawler.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/scripts/crawler.py b/scripts/crawler.py index 9bc35a3..5d898c2 100644 --- a/scripts/crawler.py +++ b/scripts/crawler.py @@ -34,7 +34,7 @@ parser.add_argument('-c', '--crawler-file', help='Crawler CSV file to log .onion sites and the amount crawled', type=str, default='crawler.csv') parser.add_argument('-b', '--blacklist-file', help='Blacklist CSV files to filter out sites with forbidden words in them', type=str, default=blcsvfile) -parser.add_argument('verified_csv', help='Input file to read for .onion links to crawl', type=str, default=verifiedcsvfile) +parser.add_argument('-V', '--verified-file', help='Input file to read for .onion links to crawl', type=str, default=verifiedcsvfile) args = parser.parse_args() script_abs_path = os.path.dirname(os.path.abspath(__file__)) @@ -77,7 +77,7 @@ def get_blacklist_file(): return None # get list of .onion links from the verified.csv file -verified_csv_file = pd.read_csv(args.verified_csv) +verified_csv_file = pd.read_csv(args.verified_file) crawler_file = get_crawler_file() output_file = get_output_file() blacklist_file = get_blacklist_file()