Update README.md
Browse files
README.md
CHANGED
|
@@ -4,7 +4,7 @@ language:
|
|
| 4 |
- en
|
| 5 |
---
|
| 6 |
|
| 7 |
-
# codegen25-7b-multi
|
| 8 |
|
| 9 |
* Model creator: [Salesforce](https://huggingface.co/Salesforce)
|
| 10 |
* Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
|
|
@@ -13,19 +13,12 @@ language:
|
|
| 13 |
|
| 14 |
This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
|
| 15 |
|
| 16 |
-
## Quantization Parameters
|
| 17 |
-
|
| 18 |
-
Weight compression was performed using `nncf.compress_weights` with the following parameters:
|
| 19 |
-
|
| 20 |
-
* mode: **INT8_ASYM**
|
| 21 |
-
|
| 22 |
-
For more information on quantization, check the [OpenVINO model optimization guide](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/weight-compression.html).
|
| 23 |
|
| 24 |
## Compatibility
|
| 25 |
|
| 26 |
The provided OpenVINO™ IR model is compatible with:
|
| 27 |
|
| 28 |
-
* OpenVINO version 2024.
|
| 29 |
* Optimum Intel 1.16.0 and higher
|
| 30 |
|
| 31 |
## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
|
|
@@ -42,7 +35,7 @@ pip install optimum[openvino] tiktoken
|
|
| 42 |
from transformers import AutoTokenizer
|
| 43 |
from optimum.intel.openvino import OVModelForCausalLM
|
| 44 |
|
| 45 |
-
model_id = "OpenVINO/codegen25-7b-multi-
|
| 46 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
| 47 |
model = OVModelForCausalLM.from_pretrained(model_id)
|
| 48 |
text = "def hello_world():"
|
|
@@ -65,8 +58,8 @@ pip install openvino-genai huggingface_hub
|
|
| 65 |
```
|
| 66 |
import huggingface_hub as hf_hub
|
| 67 |
|
| 68 |
-
model_id = "OpenVINO/codegen25-7b-multi-
|
| 69 |
-
model_path = "codegen25-7b-multi-
|
| 70 |
|
| 71 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
| 72 |
|
|
|
|
| 4 |
- en
|
| 5 |
---
|
| 6 |
|
| 7 |
+
# codegen25-7b-multi-fp16-ov
|
| 8 |
|
| 9 |
* Model creator: [Salesforce](https://huggingface.co/Salesforce)
|
| 10 |
* Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
|
|
|
|
| 13 |
|
| 14 |
This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
|
| 15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
## Compatibility
|
| 18 |
|
| 19 |
The provided OpenVINO™ IR model is compatible with:
|
| 20 |
|
| 21 |
+
* OpenVINO version 2024.2.0 and higher
|
| 22 |
* Optimum Intel 1.16.0 and higher
|
| 23 |
|
| 24 |
## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
|
|
|
|
| 35 |
from transformers import AutoTokenizer
|
| 36 |
from optimum.intel.openvino import OVModelForCausalLM
|
| 37 |
|
| 38 |
+
model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
|
| 39 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
| 40 |
model = OVModelForCausalLM.from_pretrained(model_id)
|
| 41 |
text = "def hello_world():"
|
|
|
|
| 58 |
```
|
| 59 |
import huggingface_hub as hf_hub
|
| 60 |
|
| 61 |
+
model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
|
| 62 |
+
model_path = "codegen25-7b-multi-fp16-ov"
|
| 63 |
|
| 64 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
| 65 |
|