This program downloads all files of a Stud.IP users current semester.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
Go to file
TiynGER 6d18baa8b6
adding license
4 years ago
.gitignore initial commit 4 years ago
LICENSE adding license 4 years ago
README.md adding license 4 years ago
crawler.py initial commit 4 years ago
requirements.txt initial commit 4 years ago

README.md

Stud.IP Crawler

This is a program that downloads all files available for a given Stud.IP user. It only downloads and searches through the courses in the current semester. If you run the program again it only downloads files that have changed since the last time running it.

Features/To-Dos

  • Downloads files of given users active semester via commandline
    • Keeping file structure of Stud.IP
    • Specify username
    • Specify password
    • Specify Stud.IP-URL
    • Specify output directory
    • Specify chunk size to download big files
  • Only download files after given date
    • Save and read download date
    • Possible reset of download date
  • Incremental file download
    • Indexing downloaded files and folders
  • Logging
    • Console log
    • Log file

Installation

  • git clone https://github.com/tiyn/studip-crawler
  • cd studip-crawler
  • pip3install -r requirements - install dependencies

Usage

Just run the file via python3 crawler.py [options]. Alternatively to python3 crawler.py you can give yourself permissions using chmod +x crawler.py [options] and run it with ./crawler.py [options]. There are several options required to work. Run python3 crawler.py -h for a help menu and see which ones are important for you.

Tested StudIP instances

  • Carl von Ossietzky Universität Oldenburg