Add pipeline tag and library name

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +19 -8
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
- license: mit
 
3
  datasets:
4
  - CodeGoat24/HPD
5
  - CodeGoat24/LiFT-HRA
@@ -9,11 +10,11 @@ datasets:
9
  - CodeGoat24/VideoFeedback
10
  - CodeGoat24/LLaVA-Critic-113k
11
  - CodeGoat24/VideoDPO
12
- base_model:
13
- - lmms-lab/llava-onevision-qwen2-7b-ov
 
14
  ---
15
 
16
-
17
  # Unified-Reward-7B
18
 
19
  ## Model Summary
@@ -64,12 +65,22 @@ image_tensor = [_image.to(dtype=torch.float16, device=device) for _image in imag
64
  conv_template = "qwen_1_5" # Make sure you use correct chat template for different models
65
 
66
  # pairwise ranking
67
- critic_prompt = "Given an image and a corresponding question, please serve as an unbiased and fair judge to evaluate the quality of the answers provided by a Large Multimodal Model (LMM). Determine which answer is better and explain your reasoning with specific details. Your task is provided as follows:\nQuestion: [What this image presents?]\nThe first response: [The image is a black and white sketch of a line that appears to be in the shape of a cross. The line is a simple and straightforward representation of the cross shape, with two straight lines intersecting at a point.]\nThe second response: [This is a handwritten number seven.]\nASSISTANT:\n"
 
 
 
 
 
68
 
69
  # pointwise scoring
70
- # critic_prompt = "Given an image and a corresponding question, please serve as an unbiased and fair judge to evaluate the quality of answer answers provided by a Large Multimodal Model (LMM). Score the response out of 100 and explain your reasoning with specific details. Your task is provided as follows:\nQuestion: [What this image presents?]\nThe LMM response: [This is a handwritten number seven.]\nASSISTANT:\n "
71
-
72
- question = DEFAULT_IMAGE_TOKEN + "\n" + critic_prompt
 
 
 
 
 
73
  conv = copy.deepcopy(conv_templates[conv_template])
74
  conv.append_message(conv.roles[0], question)
75
  conv.append_message(conv.roles[1], None)
 
1
  ---
2
+ base_model:
3
+ - lmms-lab/llava-onevision-qwen2-7b-ov
4
  datasets:
5
  - CodeGoat24/HPD
6
  - CodeGoat24/LiFT-HRA
 
10
  - CodeGoat24/VideoFeedback
11
  - CodeGoat24/LLaVA-Critic-113k
12
  - CodeGoat24/VideoDPO
13
+ license: mit
14
+ pipeline_tag: image-text-to-text
15
+ library_name: llava
16
  ---
17
 
 
18
  # Unified-Reward-7B
19
 
20
  ## Model Summary
 
65
  conv_template = "qwen_1_5" # Make sure you use correct chat template for different models
66
 
67
  # pairwise ranking
68
+ critic_prompt = "Given an image and a corresponding question, please serve as an unbiased and fair judge to evaluate the quality of the answers provided by a Large Multimodal Model (LMM). Determine which answer is better and explain your reasoning with specific details. Your task is provided as follows:
69
+ Question: [What this image presents?]
70
+ The first response: [The image is a black and white sketch of a line that appears to be in the shape of a cross. The line is a simple and straightforward representation of the cross shape, with two straight lines intersecting at a point.]
71
+ The second response: [This is a handwritten number seven.]
72
+ ASSISTANT:
73
+ "
74
 
75
  # pointwise scoring
76
+ # critic_prompt = "Given an image and a corresponding question, please serve as an unbiased and fair judge to evaluate the quality of answer answers provided by a Large Multimodal Model (LMM). Score the response out of 100 and explain your reasoning with specific details. Your task is provided as follows:
77
+ Question: [What this image presents?]
78
+ The LMM response: [This is a handwritten number seven.]
79
+ ASSISTANT:
80
+ "
81
+
82
+ question = DEFAULT_IMAGE_TOKEN + "
83
+ " + critic_prompt
84
  conv = copy.deepcopy(conv_templates[conv_template])
85
  conv.append_message(conv.roles[0], question)
86
  conv.append_message(conv.roles[1], None)