| | --- |
| | language: |
| | - ko |
| | base_model: |
| | - EleutherAI/polyglot-ko-1.3b |
| | pipeline_tag: text-generation |
| | tags: |
| | - ingredient |
| | - chatbot |
| | - review |
| | - usage |
| | license: apache-2.0 |
| | --- |
| | |
| | ## 🐥 Base Model |
| |
|
| | **EleutherAI/polyglot-ko-1.3b** |
| |
|
| | This model is based on the Korean version of Polyglot-1.3B, an open-source language model released by EleutherAI. |
| | It is pre-trained on a large-scale Korean corpus and designed for general-purpose Korean language understanding and generation tasks. |
| |
|
| | --- |
| |
|
| | ## Training Procedure |
| |
|
| | ### Training Hyperparameters |
| |
|
| | The following hyperparameters were used during training: |
| |
|
| | - `output_dir`: `./qlora_model_eleutherai` |
| | - `per_device_train_batch_size`: `2` |
| | - `gradient_accumulation_steps`: `4` |
| | - `total_batch_size`: `8 (2 x 4)` |
| | - `learning_rate`: `2e-5` |
| | - `num_train_epochs`: `2` |
| | - `fp16`: `True` |
| | - `logging_dir`: `./logs` |
| | - `logging_steps`: `5` |
| | - `save_steps`: `100` |
| | - `save_total_limit`: `1` |
| | - `load_best_model_at_end`: `True` |
| | - `metric_for_best_model`: `loss` |