Open-source developers all over the world are working on millions of projects: writing code & documentation, fixing & submitting bugs, and so forth. GitHub Archive is a project to record the public GitHub timeline, archive it, and make it easily accessible for further analysis.
GitHub provides 20+ event types, which range from new commits and fork events, to opening new tickets, commenting, and adding members to a project. These events are aggregated into hourly archives, which you can access with any HTTP client:
|Activity for 1/1/2015 @ 3PM UTC||
|Activity for 1/1/2015||
|Activity for all of January 2015||
Each archive contains JSON encoded events as reported by the GitHub API. You can download the raw data and apply own processing to it - e.g. write a custom aggregation script, import it into a database, and so on! An example Ruby script to download and iterate over a single archive:
For the curious, check out The Changelog episode #144 for an in-depth interview about the history of GitHub Archive, integration with BigQuery, where the project is heading, and more.
The entire GitHub Archive is also available as a public dataset on Google BigQuery: the dataset is automatically updated every hour and enables you to run arbitrary SQL-like queries over the entire dataset in seconds. To get started:
For convenience, note that there are multiple tables that you can use for your analysis:
2015tables contain all activities for each respective year.
The schema of above datasets contains distinct columns for common activity fields (see same response format), a
"payload" string field which contains the JSON encoded activity description, and
"other" string field containing all other fields.
"payload"field is different for each event type and may be updated by GitHub at any point, hence it is kept as a serialized JSON string value in BigQuery. Use the provided JSON functions (e.g. see query example above with
JSON_EXTRACT()) to extract and access data in this field.
"other"field is a JSON string which contains all other data provided but GitHub that does not match the predefined BigQuery schema - e.g. if GitHub adds a new field, it will show up in "other" until and unless the schema is extended to support it.
Note that you get 1 TB of data processed per month free of charge. In order to make best use of it, you can restrict your queries to relevant time ranges to minimize the amount of scanned data. To scan multiple tables at once, you can use table wildcards:
Changelog Nightly is the new and improved version of the daily email reports powered by the Github Archive data. These reports ship each day at 10pm CT and unearth the hottest new repos on GitHub. Alternatively, if you want something curated and less frequent, subscribe to Changelog Weekly.
Have a cool project that should be on this list? Send a pull request!