Skip to content
This repository has been archived by the owner on Jul 17, 2023. It is now read-only.

GitHub Scraper is a tool for tracking several repositories within one Google Spreadsheet, making task management and status info sharement between teammates easier.

License

IlyaFaer/GitHub-Scraper

Repository files navigation

GitHub Scraper (beta, active development paused)

GitHub Scraper automatically builds issues and PRs tables in Google Sheets documents, and periodically updates them.

  • Multirepo: track several repositories within one sheet and several sheets within one spreadsheet
  • Constructible: tweak table structure, coloring and filling functions
  • Adaptive: change your preferences and add new repos/sheets without restarting Scraper
  • Ready to go: avoid tweaking Scraper and just use completely workable examples

image

Setup
To build your tables and start tracking repositories you need:

Scraper will build tables and start tracking specified repositories. First filling can take time, but subsequent updates are faster (~80% faster), as Scraper is processing only recently updated PRs and issues. You can check filling progress in logs.txt. If any error occur, its traceback will be shown in logs.txt as well.

Structure, auto and manual filling
You can tweak table filling in fill_funcs.py, leaving some columns for manual-only use (for example "Comment"), setting ignoring and cleanup rules, sorting, coloring, etc., in any way you like.

Scraper uses config.py as a source of preferences. Before update it reloads config.py module, so you can change preferences without stoping Scraper - add new sheets, repositories, rules, etc.

PR autodetection
To make Scraper detect PRs, use GitHub keywords "Towards", "Closes", "Fixes" to make link from PRs body to the original issue. Scraper will use these links to fill "Public PR" field in the related issues.

Archive
By default, Scraper moves issues with Done status into Archive sheet and stops tracking them. This feature allows to avoid overwhelming sheets with non-active data. The behavior can be changed by setting ARCHIVE_SHEET configurations and to_be_archived filling function.

Credentials
On a first Scraper launch you'll have to authenticate on Google Sheets API (you'll see appropriate popup-window) and on GitHub (with a console). On subsequent launches Scraper will use previous credentials without asking to enter them once again.

Beta version disclaimer
Scraper is in a state of active development yet. Please, use Releases as the most stable versions, and feel free to open an issue in case of any problems.

About

GitHub Scraper is a tool for tracking several repositories within one Google Spreadsheet, making task management and status info sharement between teammates easier.

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages