Resolved conflicts + moved constant variables to conf.py and modified code as required

This commit is contained in:
doctor_dev 2025-05-30 15:56:07 +00:00
commit ce5552369d
No known key found for this signature in database
GPG key ID: F12F7F71CB84AEAA
5 changed files with 305 additions and 477 deletions

View file

@ -1,4 +1,3 @@
###TODO: importing * is bad practice should import just utils and use it like in lantern_logic.py
from utils import *
import logic.lantern_logic as lantern
from dotenv import load_dotenv
@ -52,10 +51,6 @@ def main():
print("Official participant ",line.strip() , "'s directory doesnt exist, creating it")
os.makedirs(participantdir)
print_colors("""
;
ED.
@ -97,8 +92,8 @@ def main():
if os.path.isfile(urlpath):
with open(urlpath) as f:
instance = f.read().rstrip()
if IsOnionValid(instance):
print_colors(f"[+] Instance Name: {instance}. Valid:{IsOnionValid(instance)}")
if IsOnionLinkValid(instance):
print_colors(f"[+] Instance Name: {instance}. Valid:{IsOnionLinkValid(instance)}")
break
else:
print_colors(f'[-] Invalid instance name in ~/.darknet_participant_url: {instance}',is_error=True )
@ -107,8 +102,8 @@ def main():
print_colors("[+] Instance Path doesn't exist yet")
print_colors(f"Your url will be saved here {urlpath}")
instance = input("What is your Instance domain?(ex: lantern.nowherejezfoltodf4jiyl6r56jnzintap5vyjlia7fkirfsnfizflqd.onion): ")
if IsOnionValid(instance):
print_colors(f"[+] Instance Name: {instance}. Valid: {IsUrlValid(instance)}")
if IsOnionLinkValid(instance):
print_colors(f"[+] Instance Name: {instance}. Valid: {IsOnionLinkValid(instance)}")
instancepath=rootpath+'www/participants/'+instance
else:
print_colors(f'[-] Invalid instance name in ~/.darknet_participant_url: {instance}', is_error=True )
@ -215,9 +210,9 @@ Maintenance:
while(IsCategoryValid(category) is not True):
category = input("What is the website Category? ")
# the url of the website (required) + check if its valid
url=''
while(IsUrlValid(url) is not True and IsSimpleXChatroomValid(url) is not True):
url=input("What is the website URL ? ")
url = ''
while not IsURLValid(url):
url = input("What is the website URL ? ")
# a quick description (optional) + check if its valid
desc='DEFAULT'
@ -250,7 +245,7 @@ Maintenance:
uvdf = uvdf.sort_values(by=["Category","Score"], ascending=[True,False]) # sorting categories
print_colors("[+] New row added! now writing the csv file")
else:
print("Adding new row in verified.csv since descriptioln is not empty")
print("Adding new row in verified.csv since description is not empty")
vdf.loc[-1] = newrow # adding a row
vdf = vdf.sort_values(by=["Category","Score"], ascending=[True,False]) # sorting categories
print_colors("[+] New row added! now writing the csv file")
@ -461,7 +456,7 @@ Maintenance:
value = input("What is the new name of the website? ")
vdf.at[index,'Name']=value
elif i == 3: # column URL
while(IsUrlValid(value) is not True or value == ''):
while(IsURLValid(value) is not True or value == ''):
value = input("What is the new URL of the website? ")
vdf.at[index,'URL']=value
elif i == 4: # column Sensitive
@ -507,7 +502,7 @@ Maintenance:
value = input("What is the new name of the website? ")
uvdf.at[index,'Name']=value
elif i == 3: # column URL
while(IsUrlValid(value) is not True or value == ''):
while(IsURLValid(value) is not True or value == ''):
value = input("What is the new URL of the website? ")
uvdf.at[index,'URL']=value
elif i == 4: # column Sensitive
@ -591,7 +586,7 @@ Maintenance:
case 5:
print_colors("[+] Add a new webring participant (and download their files into their directory (without trusting them yet!))")
webring_participant_url = ''
while(IsOnionValid(webring_participant_url) is not True):
while(IsOnionLinkValid(webring_participant_url) is not True):
webring_participant_url = input("What is the onion domain of the new webring participant? (ex: lantern.nowherejezfoltodf4jiyl6r56jnzintap5vyjlia7fkirfsnfizflqd.onion) ")
participantdir=rootpath+'www/participants/'+webring_participant_url
if os.path.isdir(participantdir):
@ -695,7 +690,7 @@ Maintenance:
csvdf.to_csv(csvfilepath, index=False)
### SANITY CHECK 1: Mark all the rows that have incorrect formatting for deletion###
if IsUrlValid(csvdf.at[i, 'Instance']) is False or IsCategoryValid(csvdf.at[i, 'Category']) is False or IsNameValid(csvdf.at[i, 'Name']) is False or IsUrlValid(csvdf.at[i, 'URL']) is False or IsStatusValid(csvdf.at[i, 'Sensitive']) is False or IsDescriptionValid(csvdf.at[i, 'Description']) is False or IsStatusValid(csvdf.at[i, 'Status']) is False or IsScoreValid(csvdf.at[i, 'Score']) is False:
if IsURLValid(csvdf.at[i, 'Instance']) is False or IsCategoryValid(csvdf.at[i, 'Category']) is False or IsNameValid(csvdf.at[i, 'Name']) is False or IsURLValid(csvdf.at[i, 'URL']) is False or IsStatusValid(csvdf.at[i, 'Sensitive']) is False or IsDescriptionValid(csvdf.at[i, 'Description']) is False or IsStatusValid(csvdf.at[i, 'Status']) is False or IsScoreValid(csvdf.at[i, 'Score']) is False:
#mark the row for deletion as it has invalid inputs
if i not in rows2delete:
print_colors(f"Marking row {i} for deletion, as it has invalid inputs")
@ -875,7 +870,7 @@ Maintenance:
break
else:
print_colors("[+] checking if the Word/URL is valid: ")
if IsUrlValid(word) or IsOnionValid(word) or IsDescriptionValid(word):
if IsURLValid(word) or IsDescriptionValid(word):
print_colors('[+] Word/URL is valid, adding the word into the sensitive wordlist')
newrow=[word]
print_colors(f"[+] NEWROW= {newrow}")
@ -944,7 +939,7 @@ Maintenance:
break
else:
print_colors("[+] Checking if the Word/URL is valid: ")
if IsUrlValid(word) or IsOnionValid(word) or IsDescriptionValid(word):
if IsURLValid(word) or IsDescriptionValid(word):
print_colors('[+] Word/URL is valid, adding the word into the blacklist')
newrow=[word]
print_colors(f"[+] NEWROW= {newrow}")
@ -1057,11 +1052,11 @@ Maintenance:
csvdf.to_csv(csvfilepath, index=False)
### SANITY CHECK 1: Mark all the rows that have incorrect formatting for deletion###
if IsUrlValid(csvdf.at[i, 'Instance']) is False or IsCategoryValid(csvdf.at[i, 'Category']) is False or IsNameValid(csvdf.at[i, 'Name']) is False or IsUrlValid(csvdf.at[i, 'URL']) is False or IsStatusValid(csvdf.at[i, 'Sensitive']) is False or IsDescriptionValid(csvdf.at[i, 'Description']) is False or IsStatusValid(csvdf.at[i, 'Status']) is False or IsScoreValid(csvdf.at[i, 'Score']) is False:
if IsURLValid(csvdf.at[i, 'Instance']) is False or IsCategoryValid(csvdf.at[i, 'Category']) is False or IsNameValid(csvdf.at[i, 'Name']) is False or IsURLValid(csvdf.at[i, 'URL']) is False or IsStatusValid(csvdf.at[i, 'Sensitive']) is False or IsDescriptionValid(csvdf.at[i, 'Description']) is False or IsStatusValid(csvdf.at[i, 'Status']) is False or IsScoreValid(csvdf.at[i, 'Score']) is False:
if i not in rows2delete:
print_colors(f"Marking row {i} for deletion, as it has invalid inputs")
#print_colors(f"{row}")
print(IsUrlValid(csvdf.at[i, 'Instance']), IsCategoryValid(csvdf.at[i, 'Category']), IsNameValid(csvdf.at[i, 'Name']), IsUrlValid(csvdf.at[i, 'URL']), IsStatusValid(csvdf.at[i, 'Sensitive']), IsDescriptionValid(csvdf.at[i, 'Description']), IsStatusValid(csvdf.at[i, 'Status']), IsScoreValid(csvdf.at[i, 'Score']))
print(IsURLValid(csvdf.at[i, 'Instance']), IsCategoryValid(csvdf.at[i, 'Category']), IsNameValid(csvdf.at[i, 'Name']), IsURLValid(csvdf.at[i, 'URL']), IsStatusValid(csvdf.at[i, 'Sensitive']), IsDescriptionValid(csvdf.at[i, 'Description']), IsStatusValid(csvdf.at[i, 'Status']), IsScoreValid(csvdf.at[i, 'Score']))
rows2delete.append(i)
read=input("Continue?")
@ -1163,9 +1158,6 @@ Maintenance:
print_colors("Invalid Number",is_error=True)
continue
except Exception as e:
print_colors(f'Try again {e}',is_error=True)
break
@ -1174,7 +1166,6 @@ Maintenance:
print_colors("No more submissions to review, exiting.")
break
case 12:
# review the crawled websites
try:
@ -1262,12 +1253,12 @@ Maintenance:
crawled_df.to_csv(crawled_file_abs_path, index=False)
elif number == 3:
# Delete from crawled_onion.csv
# Delete from crawled_onion.csv
crawled_df.drop(index=i,inplace=True)
crawled_df.to_csv(crawled_file_abs_path, index=False)
elif number == 4:
# Add to blacklist.csv
# Add to blacklist.csv
newrow=[link]
blacklist_df.loc[-1] = newrow # adding a row
@ -1285,15 +1276,10 @@ Maintenance:
print_colors("Invalid Number",is_error=True)
continue
except Exception as e:
print_colors(f'Try again {e}',is_error=True)
break
break
finally:
print_colors("No more crawled websites to review, exiting.")
break