| | --- |
| | license: mit |
| | task_categories: |
| | - question-answering |
| | - text2text-generation |
| | - text-generation |
| | language: |
| | - en |
| | tags: |
| | - Java |
| | - Code |
| | - UnitTest |
| | - testing |
| | pretty_name: Java Method to Test ChatML |
| | size_categories: |
| | - 100K<n<1M |
| | --- |
| | # Java Method to Test ChatML |
| |
|
| | This dataset is based on the [methods2test](https://github.com/microsoft/methods2test) dataset from Microsoft. It follows the ChatML template format: `[{'role': '', 'content': ''}, {...}]`. |
| |
|
| | Originally, methods2test contains only Java methods at different levels of granularity along with their corresponding test cases. The different focal method segmentations are illustrated here: |
| |
|
| |  |
| |
|
| | To simulate a conversation between a Java developer and an AI assistant, I introduce two key parameters: |
| |
|
| | - The prompt template |
| | - The input code choice |
| |
|
| | ## Prompt Template |
| |
|
| | The prompts simulate a user asking an AI assistant to generate a test case. For this, I used six different prompt templates, replicating real-life scenarios in which a user might ask an assistant to generate test cases: |
| |
|
| | 1. Direct request (e.g., default) |
| | 2. Educational question |
| | 3. Code review |
| | 4. Bug fixing |
| | 5. Debugging (similar but slightly different) |
| | 6. TDD Approach (Test-Driven Development) |
| |
|
| | You can find these prompts in the [prompts](./assets/prompts.py) file. |
| |
|
| | ## Input Code |
| |
|
| | The input code is selected from one of the six possibilities described in the "Focal methods segmentation" figure. To simulate real-life scenarios, I randomly select among these possibilities using the following distribution: |
| |
|
| | | Split | fm | +fc | +c | +m | +f | |
| | |-------|---- |---- |---- |---- |---- | |
| | | Train | 10% | 50% | 0% | 10% | 30% | |
| | | Test | 0% | 50% | 0% | 0% | 50% | |
| |
|
| | There is a bit more diversity in the training data to prevent underfitting during LLM fine-tuning. If you prefer, you can merge the splits and shuffle the data. |
| |
|
| | ## Summary |
| |
|
| | This dataset is a great starting point for building a powerful Java coding assistant or for complementing a larger knowledge base to train more advanced foundational models. |
| |
|
| | If you have any suggestions or spot any bugs, feel free to open a discussion! 😉 If you find it useful, don’t hesitate to give it a like. Peace! ✌️ |
| |
|
| | ## Usage and Citation |
| |
|
| | To use it with [🤗datasets](https://huggingface.co/docs/datasets/): |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | ds = load_dataset("random-long-int/Java_method2test_chatml", split="train") |
| | ``` |
| |
|
| | To use it elsewhere: |
| |
|
| | ```bibtex |
| | @misc{letoquart2025java, |
| | author = {Pierre-Louis Létoquart}, |
| | title = {Dataset: Java Method to Test ChatML}, |
| | year = {2025}, |
| | publisher = {Hugging Face}, |
| | journal = {Hugging Face Hub}, |
| | howpublished = {\url{https://huggingface.co/datasets/random-long-int/Java_method2test_chatml}}, |
| | } |
| | ``` |