|
16 | 16 | import db_processing as db
|
17 | 17 | import files_processing as fp
|
18 | 18 |
|
19 |
| -def create_report(short_domain, url, case_comment, data_array, report_info_array): |
| 19 | +def create_report(short_domain, url, case_comment, data_array, report_info_array, pagesearch_ui_mark): |
20 | 20 | try:
|
21 | 21 | ip = data_array[0]
|
22 | 22 | res = data_array[1]
|
@@ -92,14 +92,16 @@ def create_report(short_domain, url, case_comment, data_array, report_info_array
|
92 | 92 | ws['A4'] = 'SITEMAP.XML EXTRACTED?'
|
93 | 93 | ws['A5'] = 'SITEMAP.XML LINKS EXTRACTED?'
|
94 | 94 | ws['A6'] = 'DORKING STATUS'
|
95 |
| - ws['A7'] = 'REPORT CREATION TIME' |
| 95 | + ws['A7'] = 'PAGESEARCH STATUS' |
| 96 | + ws['A8'] = 'REPORT CREATION TIME' |
96 | 97 | ws['B1'] = subdomains_amount
|
97 | 98 | ws['B2'] = total_socials
|
98 | 99 | ws['B3'] = robots_txt_result
|
99 | 100 | ws['B4'] = sitemap_xml_result
|
100 | 101 | ws['B5'] = sitemap_links_status
|
101 | 102 | ws['B6'] = dorking_status
|
102 |
| - ws['B7'] = report_ctime |
| 103 | + ws['B7'] = pagesearch_ui_mark |
| 104 | + ws['B8'] = report_ctime |
103 | 105 |
|
104 | 106 | ws = wb['WHOIS']
|
105 | 107 | for col in ['1', '2', '3', '4', '5', '6', '7', '8']:
|
|
0 commit comments