The Course Scraper Module is a versatile tool designed to fetch course information and duration from various online learning platforms. It simplifies the process of finding the right resource for your learning needs by providing a unified interface for accessing course details from supported platforms.
- Google Developer Courses
- Fast.ai ML Course
- IBM - AI & Ethics Course
- Google Cloud Skill Boost: Machine Learning Engineer
- Google Cloud Skill Boost: Data Learning Engineer
- Google Cloud Skill Boost: Data Analyst
- Google Cloud Skill Boost: Generative AI
- Google Cloud Skill Boost: AD-HOC Courses
- Kaggle Learn Courses
- Deeplearning.ai Courses
To get started with the Course Scraper Module, follow these steps:
-
Clone the Repository:
Clone this GitHub repository to your local machine:
-
Navigate to the Course Scraper Module:
Change your current working directory to the course-scraper subfolder within the cloned repository:
cd learning-path-index/course-scraper
-
Install Dependencies:
Ensure you have all the required dependencies installed. You can do this using pip:
pip install -r requirements.txt
-
Run the Scraper:
Scrapers specific to each platform can be found in
course-scraper/src/scrapers
folder. Would you like to scrape courses from Kaggle Learn? Checkout the Kaggle Learn scraper README.md. How about Google Cloud Skill Boost? Checkout the GCSB scraper README.md.Generally scrapers can be run by navigating to the
course-scraper/src
folder, and runningpython -m scrapers.<course_platform>.<specific_script>
e.g
python -m scrapers.kaggle_learn.scrape_all_courses
-
View the Results:
The scraper will provide the course details and duration in a structured format. In the folder determined by
config.py
You can configure the general behaviour of all scrapers by modifying the config.py
file. This file allows you to specify:
- output location ✅
- the output format (TODO: 🚧),
and other settings.
We welcome contributions to enhance and expand the Course Scraper Module. If you'd like to contribute, please follow these guidelines:
- Fork the repository.
- Create a new branch for your feature or bug fix.
- Make your changes and ensure that the code passes all tests.
- Submit a pull request with a clear description of your changes and their purpose.
Happy learning and scraping!