mirror of
http://git.nowherejezfoltodf4jiyl6r56jnzintap5vyjlia7fkirfsnfizflqd.onion/nihilist/darknet-lantern.git
synced 2025-07-02 06:46:42 +00:00
fix stuff
This commit is contained in:
parent
e592602b82
commit
6150373c3f
2 changed files with 1501 additions and 1 deletions
|
@ -635,6 +635,7 @@ Maintenance:
|
||||||
print("[+] Removing the participant's duplicate entries... ")
|
print("[+] Removing the participant's duplicate entries... ")
|
||||||
# REMOVE DUPLICATES !!! do not accept any duplicate from remote participants
|
# REMOVE DUPLICATES !!! do not accept any duplicate from remote participants
|
||||||
csvdf = csvdf.drop_duplicates(subset=['URL'], keep="first", inplace=False)
|
csvdf = csvdf.drop_duplicates(subset=['URL'], keep="first", inplace=False)
|
||||||
|
csvdf = csvdf.drop_duplicates(subset=['Name'], keep="first", inplace=False)
|
||||||
csvdf.to_csv(csvfilepath, index=False)
|
csvdf.to_csv(csvfilepath, index=False)
|
||||||
|
|
||||||
csvdf = pd.read_csv(csvfilepath, on_bad_lines='skip')
|
csvdf = pd.read_csv(csvfilepath, on_bad_lines='skip')
|
||||||
|
@ -649,6 +650,7 @@ Maintenance:
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
################################ SANITY CHECKS ####################################
|
################################ SANITY CHECKS ####################################
|
||||||
### SANITY CHECK 0: make sure that ✔️ and x are replaced with YES/NO, as it changed since v1.0.1 ###
|
### SANITY CHECK 0: make sure that ✔️ and x are replaced with YES/NO, as it changed since v1.0.1 ###
|
||||||
if csvdf.at[i, 'Status'] == "✔️" or csvdf.at[i, 'Status'] == "YES" :
|
if csvdf.at[i, 'Status'] == "✔️" or csvdf.at[i, 'Status'] == "YES" :
|
||||||
|
@ -689,13 +691,16 @@ Maintenance:
|
||||||
filterterm=csvdf.at[i, 'URL']
|
filterterm=csvdf.at[i, 'URL']
|
||||||
#print('1)',filterterm)
|
#print('1)',filterterm)
|
||||||
filter_vdf= vdf[vdf.URL.str.contains(filterterm,na=False)]
|
filter_vdf= vdf[vdf.URL.str.contains(filterterm,na=False)]
|
||||||
|
filter_vdf2= vdf[vdf.Name.str.contains(filterterm,na=False)] # do not accept the new link if the name already exists in verified.csv
|
||||||
#print('2)',filter_vdf)
|
#print('2)',filter_vdf)
|
||||||
#print('3)',uvdf[uvdf.URL.str.contains(filterterm,na=False)] )
|
#print('3)',uvdf[uvdf.URL.str.contains(filterterm,na=False)] )
|
||||||
uvdf = pd.read_csv(unverifiedcsvfile, on_bad_lines='skip')
|
uvdf = pd.read_csv(unverifiedcsvfile, on_bad_lines='skip')
|
||||||
# TODO DELETE ALL DUPLICATES OF UVDF !
|
# TODO DELETE ALL DUPLICATES OF UVDF !
|
||||||
uvdf = uvdf.drop_duplicates(subset=['URL'], keep="first", inplace=False)
|
uvdf = uvdf.drop_duplicates(subset=['URL'], keep="first", inplace=False)
|
||||||
|
uvdf = uvdf.drop_duplicates(subset=['Name'], keep="first", inplace=False)
|
||||||
filter_uvdf= uvdf[uvdf.URL.str.contains(filterterm,na=False)]
|
filter_uvdf= uvdf[uvdf.URL.str.contains(filterterm,na=False)]
|
||||||
if len(filter_uvdf.index) == 0 and len(filter_vdf.index) == 0:
|
filter_uvdf2= uvdf[uvdf.Name.str.contains(filterterm,na=False)] # do not accept the new link if the name already exists in unverified.csv
|
||||||
|
if len(filter_uvdf.index) == 0 and len(filter_vdf.index) == 0 and len(filter_uvdf2.index) == 0 and len(filter_vdf2.index) == 0 :
|
||||||
newrow=row
|
newrow=row
|
||||||
uvdf.loc[-1] = newrow # adding a row
|
uvdf.loc[-1] = newrow # adding a row
|
||||||
uvdf.index = uvdf.index + 1 # shifting index
|
uvdf.index = uvdf.index + 1 # shifting index
|
||||||
|
@ -723,6 +728,22 @@ Maintenance:
|
||||||
|
|
||||||
#print_colors(f'[-] Rows to delete: {rows2delete}', is_error=True)
|
#print_colors(f'[-] Rows to delete: {rows2delete}', is_error=True)
|
||||||
# only delete rows after you've gone through all the unverified.csv OR verified.csv rows'
|
# only delete rows after you've gone through all the unverified.csv OR verified.csv rows'
|
||||||
|
# check for NAME duplicates and mark them for deletion:
|
||||||
|
# remove name duplicates that are in unverifie.csv yet exist in verified.csv (as verified.csv takes the priority)
|
||||||
|
if w == 'unverified.csv':
|
||||||
|
try:
|
||||||
|
# check if the given row Name already exists in verified.csv
|
||||||
|
filterterm=csvdf.at[i, 'Name']
|
||||||
|
filter_vdf= vdf[vdf.Name.str.contains(filterterm,na=False)]
|
||||||
|
print('[+] CHECKING FOR DUPLIATES: ',filterterm)
|
||||||
|
if len(filter_vdf.index) != 0:
|
||||||
|
# drop the unverified.csv row if its name already exists in verified.csv
|
||||||
|
print('[+] DUPLICATE FOUND, MARKING ROW FOR DELETION: ',row)
|
||||||
|
rows2delete.append(i) #mark the row for deletion if not already done
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
for i in rows2delete:
|
for i in rows2delete:
|
||||||
row=csvdf.loc[i,:].values.tolist()
|
row=csvdf.loc[i,:].values.tolist()
|
||||||
print_colors(f'[+] REMOVING ROW: {i}{row}')
|
print_colors(f'[+] REMOVING ROW: {i}{row}')
|
||||||
|
@ -749,6 +770,24 @@ Maintenance:
|
||||||
uvdf_updated = merged_df[uvdf.columns]
|
uvdf_updated = merged_df[uvdf.columns]
|
||||||
uvdf_updated.to_csv(unverifiedcsvfile, index=False)
|
uvdf_updated.to_csv(unverifiedcsvfile, index=False)
|
||||||
print(f'[+] Updated {no_descr_filter_count} empty description(s) in your unverified.csv found on partipant\'s {w}')
|
print(f'[+] Updated {no_descr_filter_count} empty description(s) in your unverified.csv found on partipant\'s {w}')
|
||||||
|
# remove all name duplicates from your own unverified.csv file:
|
||||||
|
for i,j in uvdf.iterrows():
|
||||||
|
row=uvdf.loc[i,:].values.tolist()
|
||||||
|
# check if the given row Name already exists in verified.csv
|
||||||
|
filterterm=uvdf.at[i, 'Name']
|
||||||
|
filter_vdf= vdf[vdf.Name.str.contains(filterterm,na=False)]
|
||||||
|
print('[+] CHECKING FOR DUPLIATES: ',filterterm)
|
||||||
|
if len(filter_vdf.index) != 0:
|
||||||
|
# drop the unverified.csv row if its name already exists in verified.csv
|
||||||
|
print('[+] DUPLICATE FOUND, MARKING ROW FOR DELETION: ',row)
|
||||||
|
rows2delete.append(i) #mark the row for deletion if not already done
|
||||||
|
for i in rows2delete:
|
||||||
|
row=uvdf.loc[i,:].values.tolist()
|
||||||
|
print_colors(f'[+] REMOVING ROW: {i}{row}')
|
||||||
|
uvdf.drop(i, inplace= True)
|
||||||
|
uvdf.to_csv(unverifiedcsvfile, index=False)
|
||||||
|
rows2delete= [] # it is an empty list at first
|
||||||
|
|
||||||
break
|
break
|
||||||
|
|
||||||
|
|
||||||
|
|
1461
scripts/lantern.py.old
Normal file
1461
scripts/lantern.py.old
Normal file
File diff suppressed because it is too large
Load diff
Loading…
Add table
Add a link
Reference in a new issue