You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
TiynGER
6d18baa8b6
|
5 years ago | |
---|---|---|
.gitignore | 5 years ago | |
LICENSE | 5 years ago | |
README.md | 5 years ago | |
crawler.py | 5 years ago | |
requirements.txt | 5 years ago |
README.md
Stud.IP Crawler
This is a program that downloads all files available for a given Stud.IP user. It only downloads and searches through the courses in the current semester. If you run the program again it only downloads files that have changed since the last time running it.
Features/To-Dos
- Downloads files of given users active semester via commandline
- Keeping file structure of Stud.IP
- Specify username
- Specify password
- Specify Stud.IP-URL
- Specify output directory
- Specify chunk size to download big files
- Only download files after given date
- Save and read download date
- Possible reset of download date
- Incremental file download
- Indexing downloaded files and folders
- Logging
- Console log
- Log file
Installation
git clone https://github.com/tiyn/studip-crawler
cd studip-crawler
pip3install -r requirements
- install dependencies
Usage
Just run the file via python3 crawler.py [options]
.
Alternatively to python3 crawler.py
you can give yourself permissions using chmod +x crawler.py [options]
and
run it with ./crawler.py [options]
.
There are several options required to work.
Run python3 crawler.py -h
for a help menu and see which ones are important for you.
Tested StudIP instances
- Carl von Ossietzky Universität Oldenburg