Skip to content

xhitz/Crawl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Crawl

Python webcrawler to automate events

This tool assumes you have the Selenium/Python3 enviroment installed already.

  • Just a Simple, Lightweight automated webcrawler.

#Tutorial

  1. Download ./chromedriver.sh to disered path of tool. Copy & Paste this in your terminal to download the latest chromedriver, this will download the lastest release and automatically extract zip.

    LATEST_VERSION=$(curl -s 
    https://chromedriver.storage.googleapis.com/LATEST_RELEASE) && 
    wget -O /tmp/chromedriver.zip 
    https://chromedriver.storage.googleapis.com/$LATEST_VERSION/chromedriver_linux64.zip
    && sudo unzip /tmp/chromedriver.zip chromedriver -d 
    /usr/local/bin/;
    
  2. Download the User-Agents.txt file into this tools directory.

  3. Place path to "User-Agents.txt in this field.

    f_name = open('User-Agents.txt', 'r')

  4. Place target website in between the " ".

    web.get("http://www.websitehere.com")

  5. Place website title in between the " ".

    assert "Web Title" in web.title

  6. Place the 'Xpath' to be clicked on/used between the ' '.

    element = web.find_element_by_xpath('')

#How To Use

Just run the script with:

python3 ./chromedriver.sh ./Crawl.py

#Thats It. =]

This is my first script in python, so be easy on me! XP

#Disclaimer

This tool is for education purposes only and is no way intended to do harm or perform illegal activities.

Releases

No releases published

Packages

No packages published

Languages