princeton-nlp/SWE-bench
Viewer β’ Updated β’ 21.5k β’ 29.6k β’ 140
How to use bug-localization/BLAZE with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("feature-extraction", model="bug-localization/BLAZE", trust_remote_code=True) # Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("bug-localization/BLAZE", trust_remote_code=True, dtype="auto")BLAZE is a transformer-based bug localization model that works across languages and software projects. It enhances source-bug alignment using dynamic chunking and hard example learning, enabling precise bug localization in unseen codebases and programming languages.
BeetleBox is the largest curated dataset for bug localization:
π₯ Available on Zenodo
π Also listed on Hugging Face Datasets: bug-localization/BeetleBox
All code, usage instructions, model files, and scripts are available via:
π BLAZE Repository & Demo (Zenodo)
Please cite the following paper if you use BLAZE or BeetleBox in your work:
@article{Chakraborty2025,
title = {BLAZE: Cross-Language and Cross-Project Bug Localization via Dynamic Chunking and Hard Example Learning},
ISSN = {2326-3881},
url = {http://dx.doi.org/10.1109/TSE.2025.3579574},
DOI = {10.1109/TSE.2025.3579574},
journal = {IEEE Transactions on Software Engineering},
publisher = {Institute of Electrical and Electronics Engineers (IEEE)},
author = {Chakraborty, Partha and Alfadel, Mahmoud and Nagappan, Meiyappan},
year = {2025},
pages = {1--14}
}
Base model
codesage/codesage-base