| --- |
| annotations_creators: |
| - author |
| license: |
| - gpl-3.0 |
| multilinguality: |
| - monolingual |
| pretty_name: GitHub-Python |
| dataset_name: github-python |
| dataset_type: code |
| tags: |
| - code |
| - python |
| size_categories: |
| - 100K<n⩽1M |
| task_categories: |
| - text-generation |
| --- |
| |
| # GitHub-Python — Licensed & Elaborated Variants |
|
|
| This repository ships **two complementary Python-code corpora** extracted from |
| public GitHub: |
|
|
| - **Licensed Subset** – strictly _permissive-licensed_ files suitable for |
| commercial redistribution / model training (main corpus used in our |
| experiments). |
| - **Elaborated Collection** – a broader crawl that additionally contains files |
| under _copyleft_ or unclear licenses (GPL/AGPL/LGPL, etc.). Useful for |
| analysis or pre-training where license mixing is acceptable. |
|
|
| Both variants target **code-completion / generation** research. |
|
|
| ## Dataset at a glance |
|
|
| | | **Licensed Subset** | **Elaborated Collection** | |
| | ------------------- | ------------------- | ------------------------- | |
| | Files (.py) | 53,017 | 186,066 | |
| | Unique repositories | 16,447 | 59,852 | |
| | Repository owners | 12,515 | 43,517 | |
| | Compressed size | 732 MB | 2.4 GB \* | |
| | Vocabulary (tokens) | 443,431 | 443,431 † | |
| | License coverage | Permissive only | Mixed (perm. + copyleft) | |
| | Secrets redacted | ✅ | ⚠️ not guaranteed | |
| | Time window | ≥ 2015-01-01 | ≥ 2015-01-01 | |
|
|
| \* estimated – elaborated corpus is distributed as raw file list, not a single |
| text file. |
| † same tokenizer file is shared by both variants. |
|
|
| Numbers were obtained from the final redacted corpus and companion metadata. |
|
|
| --- |
|
|
| ## Dataset structure |
|
|
| ``` |
| huggingface_dataset/ |
| ├─ mega_licensed_corpus_redacted.txt # Licensed Subset – concatenated code |
| ├─ python_files.txt # Licensed Subset – raw file URLs |
| ├─ python_files_elaborated.txt # Elaborated Collection – raw file URLs |
| ├─ python_files_elaborated_metadata.csv # Elaborated Collection metadata |
| └─ custom_tokens_vocab.txt # `<token>\t<id>` vocabulary file |
| ``` |
|
|
| ## Important Note |
|
|
| For technical reasons, seperate splits have been stored as seperate Dataset instances. See https://huggingface.co/datasets/jblitzar/github-python-metadata, https://huggingface.co/datasets/jblitzar/github-python-meta-elaborated, and https://huggingface.co/datasets/jblitzar/github-python-corpus . |
|
|
| ### File separator |
|
|
| Individual files are concatenated with the sentinel line: |
|
|
| ``` |
| # <FILESEP> |
| ``` |
|
|
| Anything following the sentinel until the next sentinel (or EOF) is the source |
| code of one file. |
|
|
| --- |
|
|
| ## Dataset variants |
|
|
| ### 1. Licensed Subset (`mega_licensed_corpus_redacted.txt`) |
| |
| • 53 K permissively-licensed files (MIT/BSD/Apache/ISC/Unlicense). |
| • All API keys & credentials removed. |
| • Ready for redistribution & commercial use (respect upstream NOTICE files). |
| |
| ### 2. Elaborated Collection (`python_files_elaborated.txt`) |
| |
| • 186 K files from a much larger crawl. |
| • Contains **GPL / LGPL / AGPL and other copyleft** licenses. |
| • Shipped _as URL list_ + metadata CSV; you must download the files yourself |
| (`datasets.load_dataset` streaming, `wget`, etc.). |
| • **No license filtering or secret-redaction performed** – use with caution. |
|
|
| When first loading the dataset, decide which variant aligns with your use case |
| (e.g. proprietary model training → Licensed Subset only). |
|
|
|
|
| --- |
|
|
| ## Collection methodology |
|
|
| 1. **Repository discovery** |
|
|
| - Queried GitHub REST API for projects with **≥ 10 stars** |
| (earlier iterations used 100+, later expanded for coverage). |
| - Only repositories with primary language _Python_ and last commit ≥ 2015. |
| |
| 2. **File filtering** |
|
|
| - Retain files whose **size ∈ [1 KB, 100 KB]**. |
| - Exclude common build/packaging scripts (`setup.py`, `__init__.py`, etc.). |
|
|
| 3. **License compliance** |
|
|
| - Allowed: MIT, Apache-2.0, BSD-2/3-Clause, ISC, Unlicense. |
| - GPL, LGPL, AGPL and proprietary licenses were **excluded**. |
|
|
| 4. **Deduplication** |
|
|
| - Unique file SHA hashes; duplicates skipped. |
|
|
| 5. **Formatting & cleaning** |
|
|
| - Formatted with _autopep8_ to normalise whitespace. |
| - Custom script removed trailing whitespace & normalised newlines. |
|
|
| 6. **Secret redaction** |
| - `truffleHog` + custom regex pass removed >150 active credentials. |
| - Redacted corpus stored as `mega_licensed_corpus_redacted.txt`. |
|
|
| --- |
|
|
| ## Custom tokenisation |
|
|
| The accompanying `custom_tokens_vocab.txt` implements a **Python-aware |
| sub-token scheme**: |
|
|
| 1. Strip doc-strings & comments. |
| 2. Split on: |
| - Camel-Case boundaries (`Camel` → `Camel`, `Case`) |
| - Underscores, spaces |
| - Indentation & newlines (preserved as `<newline>` token) |
| 3. Rare tokens (frequency < 10) were dropped → 443 k vocabulary. |
|
|
| Example: |
|
|
| ```python |
| def helloWorld(value): |
| return value + 1 |
| ``` |
|
|
| tokenises to: |
|
|
| ``` |
| def hello world ( value ) <newline> <tab> return value + 1 <newline> |
| ``` |
|
|
| --- |
|
|
| ## Usage |
|
|
| ```python |
| from datasets import load_dataset |
| |
| ds = load_dataset("jblitzar/github-python-corpus", split="train") |
| |
| print(ds[0]["code"][:300]) # raw source code |
| ``` |
|
|
| If you prefer token level examples (small reasons: memory), map the tokenizer: |
|
|
| ```python |
| from tokenizers import Tokenizer |
| tok = Tokenizer.from_file("custom_tokens_vocab.txt") |
| |
| def encode(ex): |
| ex["input_ids"] = tok.encode(ex["code"]).ids |
| return ex |
| |
| ds = ds.map(encode, remove_columns=["code"]) |
| ``` |
|
|
| --- |
|
|
| ## Ethical considerations & limitations |
|
|
| - **Licenses respected** – only permissive licenses included; retain NOTICE |
| files when redistributing derivative works. |
| - **Secrets removed** – automated & manual audits performed, yet users **must |
| not assume zero secrets**; re-audit before public deployments. |
| - **Code quality** – projects vary in style & correctness. Generated models |
| may replicate bugs or vulnerable patterns. |
|
|
| --- |
|
|
| ## Citation |
|
|
| If you use this dataset, please cite: |
|
|
| ``` |
| @misc{github-python-2024, |
| author = {JBlitzar}, |
| title = {GitHub-Python: A Permissively Licensed Corpus of Python Code}, |
| year = {2024}, |
| howpublished = {\url{https://huggingface.co/datasets/jblitzar/github-python}}, |
| note = {Version 1.0} |
| } |
| ``` |
|
|
| --- |
|
|
| ## License |
|
|
| Dataset card and aggregation scripts: **GPLv3**. |
| Each code snippet remains under its **original repository license** (MIT, |
| Apache-2.0, BSD, ISC, etc.). Users must comply with upstream notices when |
| redistributing code or derivatives. |
|
|
| <!--bump--> |