Skip to content

Extracted various fields of information from housing website using web scraping. As the website requires redirecting to different url and scrolling and many other tasks, Selenium is used in order to facilitate these tasks. There are 3 parameters required for the scroll function, they are driver, timeout (time to wait till next scroll) and number…

Notifications You must be signed in to change notification settings

AakashBelide/web_scraping-Internship_task

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

web_scraping-Intership_task

Extracted various fields of information from housing website using web scraping. As the website requires redirecting to different url and scrolling and many other tasks, Selenium is used in order to facilitate these tasks. There are 3 parameters required for the scroll function, they are driver, timeout (time to wait till next scroll) and number of times to scroll respectively. Set the number of times to scroll according to the number of data entries you require. Magicbricks loads 30 new houses everytime you scroll and reach the end of the website, so as I require information of 500 houses, I have set it as 18 (18 x30 = 540), and later I have kept 500 in the for loop for only information of 500 houses in the csv file.

Python libraries used:

Requirements:

Directions to use:

  • This is designed for only magicbricks. You can provide the searched url in a specific locality.
  • Make sure you have good internet connection and keep the system on while the process takes place.
  • You will get various pieces of information from the website like:
  • Price
  • Headline
  • Address
  • Owner
  • Bedrooms
  • Bathrooms
  • Balconies
  • Super area
  • Price Per sqft
  • Status
  • Transaction type
  • Floor
  • Car parking
  • Furnished
  • Lifts
  • Descrption
  • Price breakup
  • Address2
  • Landmarks
  • Age of Construction
  • Price comparison
  • Expected rent
  • Monthly EMI

About

Extracted various fields of information from housing website using web scraping. As the website requires redirecting to different url and scrolling and many other tasks, Selenium is used in order to facilitate these tasks. There are 3 parameters required for the scroll function, they are driver, timeout (time to wait till next scroll) and number…

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages