| --- |
| license: other |
| license_name: deepcode-ai |
| license_link: LICENSE |
| --- |
| |
| ## 3. Quick Start |
| ### Installation |
| On the basis of `Python >= 3.8` environment, install the necessary dependencies by running the following command: |
| ```shell |
| git clone https://github.com/deepcode-ai/DeepCode-VL |
| cd DeepCode-VL |
| pip install -e . |
| ``` |
| ### Simple Inference Example |
| ```python |
| import torch |
| from transformers import AutoModelForCausalLM |
| from deepcode_vl.models import VLChatProcessor, MultiModalityCausalLM |
| from deepcode_vl.utils.io import load_pil_images |
| # specify the path to the model |
| model_path = "deepcode-ai/deepcode-base" |
| vl_chat_processor: VLChatProcessor = VLChatProcessor.from_pretrained(model_path) |
| tokenizer = vl_chat_processor.tokenizer |
| vl_gpt: MultiModalityCausalLM = AutoModelForCausalLM.from_pretrained(model_path, trust_remote_code=True) |
| vl_gpt = vl_gpt.to(torch.bfloat16).cuda().eval() |
| conversation = [ |
| { |
| "role": "User", |
| "content": "<image_placeholder>Describe each stage of this image.", |
| "images": ["./images/training_pipelines.png"] |
| }, |
| { |
| "role": "Assistant", |
| "content": "" |
| } |
| ] |
| # load images and prepare for inputs |
| pil_images = load_pil_images(conversation) |
| prepare_inputs = vl_chat_processor( |
| conversations=conversation, |
| images=pil_images, |
| force_batchify=True |
| ).to(vl_gpt.device) |
| # run image encoder to get the image embeddings |
| inputs_embeds = vl_gpt.prepare_inputs_embeds(**prepare_inputs) |
| # run the model to get the response |
| outputs = vl_gpt.language_model.generate( |
| inputs_embeds=inputs_embeds, |
| attention_mask=prepare_inputs.attention_mask, |
| pad_token_id=tokenizer.eos_token_id, |
| bos_token_id=tokenizer.bos_token_id, |
| eos_token_id=tokenizer.eos_token_id, |
| max_new_tokens=512, |
| do_sample=False, |
| use_cache=True |
| ) |
| answer = tokenizer.decode(outputs[0].cpu().tolist(), skip_special_tokens=True) |
| print(f"{prepare_inputs['sft_format'][0]}", answer) |
| ``` |
| ### CLI Chat |
| ```bash |
| python cli_chat.py --model_path "deepcode-ai/deepcode-base" |
| # or local path |
| python cli_chat.py --model_path "local model path" |
| `` |