ZhangYuanhan commited on
Commit
044f7ff
1 Parent(s): 091ecd3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -10,7 +10,7 @@ metrics:
10
  tags:
11
  - multimodal
12
  model-index:
13
- - name: LLaVA-NeXT-Video-7B-Qwen2
14
  results:
15
  - task:
16
  type: multimodal
@@ -116,7 +116,7 @@ base_model:
116
  - lmms-lab/llava-onevision-qwen2-7b-si
117
  ---
118
 
119
- # LLaVA-NeXT-Video-7B-Qwen2-Video-Only
120
 
121
  ## Table of Contents
122
 
@@ -132,7 +132,7 @@ base_model:
132
  In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
133
 
134
 
135
- This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-NeXT-Video-7B-Qwen2 in terms of video benchmarks.
136
 
137
  - **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
138
  - **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
@@ -184,7 +184,7 @@ def load_video(self, video_path, max_frames_num,fps=1,force_sample=False):
184
  spare_frames = vr.get_batch(frame_idx).asnumpy()
185
  # import pdb;pdb.set_trace()
186
  return spare_frames,frame_time,video_time
187
- pretrained = "lmms-lab/LLaVA-NeXT-Video-7B-Qwen2-Video-Only "
188
  model_name = "llava_qwen"
189
  device = "cuda"
190
  device_map = "auto"
 
10
  tags:
11
  - multimodal
12
  model-index:
13
+ - name: LLaVA-Video-7B-Qwen2
14
  results:
15
  - task:
16
  type: multimodal
 
116
  - lmms-lab/llava-onevision-qwen2-7b-si
117
  ---
118
 
119
+ # LLaVA-Video-7B-Qwen2-Video-Only
120
 
121
  ## Table of Contents
122
 
 
132
  In contrast to lmms-lab/LLaVA-NeXT-Video-7B-Qwen2, this is a 7B model trained on [LLaVA-Video-178K](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Video-SFT-Data) only, based on Qwen2 language model with a context window of 32K tokens.
133
 
134
 
135
+ This model supports up to 110 frames and achieves comparable results to those of lmms-lab/LLaVA-Video-7B-Qwen2 in terms of video benchmarks.
136
 
137
  - **Repository:** [LLaVA-VL/LLaVA-NeXT](https://github.com/LLaVA-VL/LLaVA-NeXT?tab=readme-ov-file)
138
  - **Point of Contact:** [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/)
 
184
  spare_frames = vr.get_batch(frame_idx).asnumpy()
185
  # import pdb;pdb.set_trace()
186
  return spare_frames,frame_time,video_time
187
+ pretrained = "lmms-lab/LLaVA-Video-7B-Qwen2-Video-Only "
188
  model_name = "llava_qwen"
189
  device = "cuda"
190
  device_map = "auto"