recodehive / Scrape-ML

For new data generation Semi-supervised-sequence-learning-Project we have writtern a python script to fetchπŸ“Š, data from the πŸ’», imdb website 🌐 and converted into txt files.
https://scrape-ml.streamlit.app/
MIT License
85 stars 116 forks source link

[Feature]Automated Web Scraping and NLP Analysis Tool #224

Closed litesh1123 closed 1 month ago

litesh1123 commented 1 month ago

Describe the bug I have developed a script that takes a CSV file with website URLs, automatically scrapes data from each site without human intervention, and performs NLP analysis on the extracted data. The analysis includes parameters like positive and negative scores, total word count, and more. The results are then filled into an output CSV file. This script can scrape around 100 websites in approximately 15 minutes.

To Reproduce Steps to reproduce the behavior:

  1. Go to '...'
  2. Click on '....'
  3. Scroll down to '....'
  4. See error

Expected behavior A clear and concise description of what you expected to happen.

Screenshots this is drive link video of how this project functions https://drive.google.com/file/d/1St7s5Ei3P_3UJXkotZF2o-ats0TeC-ue/view?usp=sharing

how output would look like: Screenshot 2024-07-29 023838

Desktop (please complete the following information):

Smartphone (please complete the following information):

Additional context it can manually scrape data from bulk websites automatically and also perform import nlp analysis to understand maret scenario and conditions

github-actions[bot] commented 1 month ago

Thank you for creating this issue! πŸŽ‰ We'll look into it as soon as possible. In the meantime, please make sure to provide all the necessary details and context. If you have any questions or additional information, feel free to add them here. Your contributions are highly appreciated! 😊

You can also check our CONTRIBUTING.md for guidelines on contributing to this project.

github-actions[bot] commented 1 month ago

Hello @litesh1123! Your issue #224 has been closed. Thank you for your contribution!