Welcome to our web scrapping repository! This repository contains various scraping scripts using modules like lxml, BeautifulSoup, and Selenium to gather data from different websites. π
The repository follows a structured approach for organizing scraping scripts:
π¦ web-scrapper-repo β£ π site1 β β£ π script.py β β π README.md β β π requirements.txt β£ π site2 β β£ π script.py β β π README.md β β π requirements.txt β ...
Each folder (site1
, site2
, etc.) corresponds to a different website that we have scraped. Inside each folder, you'll find a scraping script (script.py
) and a README.md file providing information about the scraped data.
We appreciate your contributions to make this repository better! Here's how you can contribute:
If you encounter issues with the existing code, or if you have suggestions for improvements, please create an issue. Be sure to provide details and context about the problem or enhancement you're proposing.
To start contributing, fork this repository to your GitHub account, and then clone it to your local machine.
git clone https://github.com/anupammaurya6767/web_scrapper.git
Before making changes, create a new branch for your feature or bug fix.
git checkout -b feature-name
Make your changes in the appropriate folder (e.g., site1). Update the scraping script and the README.md file to include information about the data you've scraped.
Last Updated: [Date]
git add .
git commit -m "Added data from Site1"
git push origin feature-name
Once you've made your changes and pushed them to your forked repository, create a pull request to merge your changes into the main repository. Provide a clear description of your changes.
We'll review your contribution and merge it if it aligns with the repository's goals.
Thank you for contributing to our Web Scrapper Repository! π