Skip to content

Commit

Permalink
Merge pull request #46 from dmdhrumilmistry/fix-web-crawlers-in-UI
Browse files Browse the repository at this point in the history
fix dir and subdom crawler options in UI
  • Loading branch information
dmdhrumilmistry authored Apr 29, 2023
2 parents af268d2 + 5b48765 commit 515432b
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 6 deletions.
4 changes: 2 additions & 2 deletions pyhtools/UI/functions.py
Original file line number Diff line number Diff line change
Expand Up @@ -167,10 +167,10 @@ async def run():
await attacker.webspider()

elif cmd == 'webcrawldirs':
attacker.webcrawldirs()
await attacker.webcrawldirs()

elif cmd == 'webcrawlsubdom':
attacker.webcrawlsubdom()
await attacker.webcrawlsubdom()

elif cmd == 'weblogin':
attacker.brute_login()
Expand Down
8 changes: 4 additions & 4 deletions pyhtools/attackers/attackers.py
Original file line number Diff line number Diff line change
Expand Up @@ -129,23 +129,23 @@ async def webspider():
print(f'[*] Total Links Found: {len(discovered_links)}')


def webcrawldirs():
async def webcrawldirs():
'''
description: find valid directories of the website using a wordlist
params: None
returns: None
'''
domain = input('[+] DOMAIN (duckduckgo.com): ')
wordlist_path = input('[+] WORDLIST PATH: ')
run(discoverer.check_dirs(domain=domain, wordlist=wordlist_path))
await discoverer.check_dirs(domain=domain, wordlist_path=wordlist_path)


def webcrawlsubdom():
async def webcrawlsubdom():
'''
description: find valid subdomains of the website using a wordlist
params: None
returns: None
'''
domain = input('[+] DOMAIN (duckduckgo.com) : ')
wordlist_path = input('[+] WORDLIST PATH : ')
run(discoverer.check_subdomains(domain=domain, wordlist=wordlist_path))
await discoverer.check_subdomains(domain=domain, wordlist_path=wordlist_path)

0 comments on commit 515432b

Please sign in to comment.