github-python / README.md
jblitzar's picture
Update README.md
a99387c verified
---
annotations_creators:
- author
license:
- gpl-3.0
multilinguality:
- monolingual
pretty_name: GitHub-Python
dataset_name: github-python
dataset_type: code
tags:
- code
- python
size_categories:
- 100K<n⩽1M
task_categories:
- text-generation
---
# GitHub-Python — Licensed & Elaborated Variants
This repository ships **two complementary Python-code corpora** extracted from
public GitHub:
- **Licensed Subset** – strictly _permissive-licensed_ files suitable for
commercial redistribution / model training (main corpus used in our
experiments).
- **Elaborated Collection** – a broader crawl that additionally contains files
under _copyleft_ or unclear licenses (GPL/AGPL/LGPL, etc.). Useful for
analysis or pre-training where license mixing is acceptable.
Both variants target **code-completion / generation** research.
## Dataset at a glance
| | **Licensed Subset** | **Elaborated Collection** |
| ------------------- | ------------------- | ------------------------- |
| Files (.py) | 53,017 | 186,066 |
| Unique repositories | 16,447 | 59,852 |
| Repository owners | 12,515 | 43,517 |
| Compressed size | 732 MB | 2.4 GB \* |
| Vocabulary (tokens) | 443,431 | 443,431 † |
| License coverage | Permissive only | Mixed (perm. + copyleft) |
| Secrets redacted | ✅ | ⚠️ not guaranteed |
| Time window | ≥ 2015-01-01 | ≥ 2015-01-01 |
\* estimated – elaborated corpus is distributed as raw file list, not a single
text file.
† same tokenizer file is shared by both variants.
Numbers were obtained from the final redacted corpus and companion metadata.
---
## Dataset structure
```
huggingface_dataset/
├─ mega_licensed_corpus_redacted.txt # Licensed Subset – concatenated code
├─ python_files.txt # Licensed Subset – raw file URLs
├─ python_files_elaborated.txt # Elaborated Collection – raw file URLs
├─ python_files_elaborated_metadata.csv # Elaborated Collection metadata
└─ custom_tokens_vocab.txt # `<token>\t<id>` vocabulary file
```
## Important Note
For technical reasons, seperate splits have been stored as seperate Dataset instances. See https://huggingface.co/datasets/jblitzar/github-python-metadata, https://huggingface.co/datasets/jblitzar/github-python-meta-elaborated, and https://huggingface.co/datasets/jblitzar/github-python-corpus .
### File separator
Individual files are concatenated with the sentinel line:
```
# <FILESEP>
```
Anything following the sentinel until the next sentinel (or EOF) is the source
code of one file.
---
## Dataset variants
### 1. Licensed Subset (`mega_licensed_corpus_redacted.txt`)
• 53 K permissively-licensed files (MIT/BSD/Apache/ISC/Unlicense).
• All API keys & credentials removed.
• Ready for redistribution & commercial use (respect upstream NOTICE files).
### 2. Elaborated Collection (`python_files_elaborated.txt`)
• 186 K files from a much larger crawl.
• Contains **GPL / LGPL / AGPL and other copyleft** licenses.
• Shipped _as URL list_ + metadata CSV; you must download the files yourself
(`datasets.load_dataset` streaming, `wget`, etc.).
**No license filtering or secret-redaction performed** – use with caution.
When first loading the dataset, decide which variant aligns with your use case
(e.g. proprietary model training → Licensed Subset only).
---
## Collection methodology
1. **Repository discovery**
- Queried GitHub REST API for projects with **≥ 10 stars**
(earlier iterations used 100+, later expanded for coverage).
- Only repositories with primary language _Python_ and last commit ≥ 2015.
2. **File filtering**
- Retain files whose **size ∈ [1 KB, 100 KB]**.
- Exclude common build/packaging scripts (`setup.py`, `__init__.py`, etc.).
3. **License compliance**
- Allowed: MIT, Apache-2.0, BSD-2/3-Clause, ISC, Unlicense.
- GPL, LGPL, AGPL and proprietary licenses were **excluded**.
4. **Deduplication**
- Unique file SHA hashes; duplicates skipped.
5. **Formatting & cleaning**
- Formatted with _autopep8_ to normalise whitespace.
- Custom script removed trailing whitespace & normalised newlines.
6. **Secret redaction**
- `truffleHog` + custom regex pass removed >150 active credentials.
- Redacted corpus stored as `mega_licensed_corpus_redacted.txt`.
---
## Custom tokenisation
The accompanying `custom_tokens_vocab.txt` implements a **Python-aware
sub-token scheme**:
1. Strip doc-strings & comments.
2. Split on:
- Camel-Case boundaries (`Camel``Camel`, `Case`)
- Underscores, spaces
- Indentation & newlines (preserved as `<newline>` token)
3. Rare tokens (frequency < 10) were dropped → 443 k vocabulary.
Example:
```python
def helloWorld(value):
return value + 1
```
tokenises to:
```
def hello world ( value ) <newline> <tab> return value + 1 <newline>
```
---
## Usage
```python
from datasets import load_dataset
ds = load_dataset("jblitzar/github-python-corpus", split="train")
print(ds[0]["code"][:300]) # raw source code
```
If you prefer token level examples (small reasons: memory), map the tokenizer:
```python
from tokenizers import Tokenizer
tok = Tokenizer.from_file("custom_tokens_vocab.txt")
def encode(ex):
ex["input_ids"] = tok.encode(ex["code"]).ids
return ex
ds = ds.map(encode, remove_columns=["code"])
```
---
## Ethical considerations & limitations
- **Licenses respected** – only permissive licenses included; retain NOTICE
files when redistributing derivative works.
- **Secrets removed** – automated & manual audits performed, yet users **must
not assume zero secrets**; re-audit before public deployments.
- **Code quality** – projects vary in style & correctness. Generated models
may replicate bugs or vulnerable patterns.
---
## Citation
If you use this dataset, please cite:
```
@misc{github-python-2024,
author = {JBlitzar},
title = {GitHub-Python: A Permissively Licensed Corpus of Python Code},
year = {2024},
howpublished = {\url{https://huggingface.co/datasets/jblitzar/github-python}},
note = {Version 1.0}
}
```
---
## License
Dataset card and aggregation scripts: **GPLv3**.
Each code snippet remains under its **original repository license** (MIT,
Apache-2.0, BSD, ISC, etc.). Users must comply with upstream notices when
redistributing code or derivatives.
<!--bump-->