From 6cc297fe66874d29be57a97ffcb8c1d2aa12f8c1 Mon Sep 17 00:00:00 2001 From: cynthia Date: Fri, 4 Apr 2025 18:10:35 +0000 Subject: [PATCH] add dotenv --- scripts/crawler.py | 15 +++++++++++++-- 1 file changed, 13 insertions(+), 2 deletions(-) diff --git a/scripts/crawler.py b/scripts/crawler.py index fcdf416..b4cc6d0 100644 --- a/scripts/crawler.py +++ b/scripts/crawler.py @@ -7,6 +7,8 @@ import os import re from utils import print_colors, IsUrlValid +from dotenv import load_dotenv + parser = argparse.ArgumentParser( prog='Lantern crawler', description='Crawls .onion sites for links to more .onion sites') @@ -20,12 +22,21 @@ parser.add_argument('-c', '--crawler-file', parser.add_argument('verified_csv', help='Input file to read for .onion links to crawl', type=str, default='verified.csv') args = parser.parse_args() +script_abs_path = os.path.dirname(os.path.abspath(__file__)) +env_path = os.path.join(script_abs_path+"/.env") +default_env_path = os.path.join(script_abs_path+"/.env.sample") + +if os.path.exists(env_path): + load_dotenv(dotenv_path=env_path) +else: + load_dotenv(dotenv_path=default_env_path) + tor_host = os.getenv("TOR_HOST") tor_port = os.getenv("TOR_PORT") session = requests.session() session.proxies = { - 'http': f'socks5h://{tor_host}:{tor_port}', - 'https': f'socks5h://{tor_host}:{tor_port}' + 'http': f'{tor_host}:{tor_port}', + 'https': f'{tor_host}:{tor_port}' } # Set user agent too for the crawler session.headers.update({'User-Agent': 'LanternCrawler'})