ML Nexus is an open-source collection of machine learning projects, covering topics like neural networks, computer vision, and NLP. Whether you're a beginner or expert, contribute, collaborate, and grow together in the world of AI. Join us to shape the future of machine learning!
Is your feature request related to a problem? Please describe.
The LinkedIn Job Scraper is a Python-based web scraping tool that uses Scrapy, a powerful web crawling framework, to extract job listings from LinkedIn. The tool is designed to circumvent common anti-scraping mechanisms like IP blocking by integrating proxy rotation, which ensures that the scraper can maintain a steady scraping speed while avoiding detection or throttling.
Key Features:
Scrapes job listings (job title, company, location, posting date, and description).
Uses proxies for IP rotation to avoid detection.
Allows the user to filter job results by keywords, locations, and other criteria.
Stores the scraped data in a structured format (e.g., CSV, JSON)
Thanks for creating the issue in ML-Nexus!π
Before you start working on your PR, please make sure to:
β Star the repository if you haven't already.
Pull the latest changes to avoid any merge conflicts.
Attach before & after screenshots in your PR for clarity.
Include the issue number in your PR description for better tracking.
Don't forget to follow @UppuluriKalyani β Project Admin β for more updates!
Tag @Neilblaze,@SaiNivedh26 for assigning the issue to you.
Happy open-source contributing!βΊοΈ
Is your feature request related to a problem? Please describe. The LinkedIn Job Scraper is a Python-based web scraping tool that uses Scrapy, a powerful web crawling framework, to extract job listings from LinkedIn. The tool is designed to circumvent common anti-scraping mechanisms like IP blocking by integrating proxy rotation, which ensures that the scraper can maintain a steady scraping speed while avoiding detection or throttling.
Key Features:
Scrapes job listings (job title, company, location, posting date, and description). Uses proxies for IP rotation to avoid detection. Allows the user to filter job results by keywords, locations, and other criteria. Stores the scraped data in a structured format (e.g., CSV, JSON)