| --- |
| dataset_info: |
| features: |
| - name: id |
| dtype: int64 |
| - name: input |
| list: |
| - name: type |
| dtype: string |
| - name: content |
| dtype: string |
| - name: output |
| struct: |
| - name: veo3 |
| list: string |
| - name: framepack |
| list: string |
| - name: framepack_seleted_video |
| dtype: string |
| - name: hunyuan |
| list: string |
| - name: hunyuan_seleted_video |
| dtype: string |
| - name: wan2.2-14b |
| list: string |
| - name: wan2.2-14b_seleted_video |
| dtype: string |
| - name: wan2.2-5b |
| list: string |
| - name: wan2.2-5b_seleted_video |
| dtype: string |
| splits: |
| - name: train |
| num_bytes: 98746 |
| num_examples: 99 |
| download_size: 36034 |
| dataset_size: 98746 |
| configs: |
| - config_name: default |
| data_files: |
| - split: train |
| path: data/train-* |
| --- |
| |
| # Visual-Intelligence |
|
|
| ## π Links |
|
|
| - [πΎ Github Repo](https://github.com/Entroplay/Visual-Intelligence) |
| - [π€ HF Dataset](https://huggingface.co/datasets/Entroplay/Visual-Intelligence) |
| - [π Blog](https://entroplay.ai/research/video-intelligence) |
|
|
| ## π Dataset Introduction |
|
|
| ### Dataset Schema |
|
|
| - **id**: Unique sample identifier. |
| - **input**: Ordered list describing the input context. |
| - **type**: Either "image" or "text". |
| - **content**: For "image", a relative path to the first-frame image. For "text", the prompt text. |
| - **output**: Generated candidates and final selections by model. |
| - **veo3**: Relative paths to videos generated by the VEO3 pipeline. |
| - **framepack**: Relative paths to videos generated by FramePack across multiple runs. |
| - **hunyuan**: Relative paths to videos generated by Hunyuan across multiple runs. |
| - **wan2.2-5b**: Relative paths to videos generated by Wan-2.2-5B across multiple runs. |
| - **wan2.2-14b**: Relative paths to videos generated by Wan-2.2-14B across multiple runs. |
| - **framepack_seleted_video**: Selected best video among FramePack candidates. |
| - **hunyuan_seleted_video**: Selected best video among Hunyuan candidates. |
| - **wan2.2-5b_seleted_video**: Selected best video among Wan 2.2 5B candidates. |
| - **wan2.2-14b_seleted_video**: Selected best video among Wan 2.2 14B candidates. |
|
|
| ### Data Format: |
|
|
| ```json |
| { |
| "id": 1, |
| "input": [ |
| { "type": "image", "content": "thumbnails/mp4/keypoint_localization.jpg" }, |
| { "type": "text", "content": "Add a bright blue dot at the tip of the branch on which the macaw is sitting. ..." } |
| ], |
| "output": { |
| "veo3": ["videos/mp4/keypoint_localization.mp4"], |
| "framepack": [ |
| "videos/1_framepack_1.mp4", |
| "videos/1_framepack_2.mp4" |
| ], |
| "hunyuan": [ |
| "videos/1_hunyuan_1.mp4", |
| "videos/1_hunyuan_2.mp4" |
| ], |
| "wan2.2-5b": [ |
| "videos/1_wan2.2-5b_1.mp4", |
| "videos/1_wan2.2-5b_2.mp4" |
| ], |
| "wan2.2-14b": [ |
| "videos/1_wan2.2-14b_1.mp4", |
| "videos/1_wan2.2-14b_2.mp4" |
| ], |
| "framepack_seleted_video": "videos/1_framepack_1.mp4", |
| "hunyuan_seleted_video": "videos/1_hunyuan_1.mp4", |
| "wan2.2-5b_seleted_video": "videos/1_wan2.2-5b_1.mp4", |
| "wan2.2-14b_seleted_video": "videos/1_wan2.2-14b_1.mp4" |
| } |
| } |
| ``` |
|
|
| ## π About project |
|
|
| Google' Veo 3 shows extreme promise in visual intelligence, demonstrating strong visual commonsense and reasoning in visual generation. We aim to construct a fully open-source evaluation suite to measure current progress in video generative intelligence across various dimensions among several state-of-the-art proprietary and open-source models. |