Skip to content

NhanAZ-Web/github-repo-crawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

GitHub Repo Crawler (CSV Export)

Tool for exporting all public repositories of a GitHub user or organization to CSV, running entirely in your browser.

What this does

  • Fetches all public repositories of a given user or organization via the GitHub REST API v3.
  • Collects repository metadata (stars, forks, issues, pull requests, contributors, releases, languages, etc.).
  • Generates and automatically downloads a CSV file with one row per repository.

How to use

  1. Open index.html directly in a modern browser (double‑click or drag into the browser; no server required).
  2. (Recommended) Paste a GitHub Personal Access Token in the token field.
    • Without a token, the GitHub API limits you to 60 unauthenticated requests per hour.
  3. Enter the GitHub username or organization name you want to analyze.
  4. Click “Fetch & Export CSV”.
  5. Wait while progress messages appear below the button. When finished, a CSV file named like
    USERNAME_github_repos_YYYY-MM-DD.csv will be downloaded automatically.

Notes

  • All work happens in the browser; data is only sent to api.github.com.
  • The tool respects GitHub rate limits and may pause briefly when close to the limit.
  • Only public repositories are included.

About

Tool for exporting all public repositories of a GitHub user or organization to CSV.

Resources

Stars

Watchers

Forks

Contributors