FineVision
Pillar scores are computed during the next indexing cycle.
--- dataset_info: - config_name: CoSyn_400k_chart features: - name: images list: image - name: texts list: - name: user dtype: string - name: assistant dtype: string - name: source dtype: string - name: relevance_ratings list: int64 - name: relevance_min dtype: int64 - name: visual_dependency_ratings list: int64 - name: visual_dependency_min dtype: int64 - name: image_correspondence_ratings list: int64 - name: image_correspondence_min dtype: int64 - name: formatting_ratings list: int64 - name...
| Entity Passport | |
| Registry ID | hf-dataset--huggingfacem4--finevision |
| Provider | huggingface |
Cite this dataset
Academic & Research Attribution
@misc{hf_dataset__huggingfacem4__finevision,
author = {HuggingFaceM4},
title = {FineVision Dataset},
year = {2026},
howpublished = {\url{https://huggingface.co/datasets/HuggingFaceM4/FineVision}},
note = {Accessed via Free2AITools Knowledge Fortress}
} đŦTechnical Deep Dive
Full Specifications [+]âž
âī¸ Nexus Index V2.0
đŦ Index Insight
FNI V2.0 for FineVision: Semantic (S:50), Authority (A:0), Popularity (P:0), Recency (R:0), Quality (Q:0).
Verification Authority
đī¸ Data Preview
Row-level preview not available for this dataset.
Schema structure is shown in the Field Logic panel when available.
đ Explore Full Dataset âđ§Ŧ Field Logic
Schema not yet indexed for this dataset.
Dataset Specification
dataset_info:
- config_name: CoSyn_400k_chart
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 25619852113.664
num_examples: 116814
download_size: 25239736178
dataset_size: 25619852113.664
- name: images
- config_name: CoSyn_400k_chemical
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 284197936.992
num_examples: 8942
download_size: 273097193
dataset_size: 284197936.992
- name: images
- config_name: CoSyn_400k_circuit
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 395788840.72
num_examples: 10470
download_size: 381928378
dataset_size: 395788840.72
- name: images
- config_name: CoSyn_400k_diagram
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 7305286810.288
num_examples: 34963
download_size: 7234372451
dataset_size: 7305286810.288
- name: images
- config_name: CoSyn_400k_document
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 24180793229.832
num_examples: 71282
download_size: 24015209877
dataset_size: 24180793229.832
- name: images
- config_name: CoSyn_400k_graphic
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 335694057.168
num_examples: 26968
download_size: 313408282
dataset_size: 335694057.168
- name: images
- config_name: CoSyn_400k_math
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 6107895469.064
num_examples: 66714
download_size: 6057279470
dataset_size: 6107895469.064
- name: images
- config_name: CoSyn_400k_music
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 405064954.03
num_examples: 11969
download_size: 377684174
dataset_size: 405064954.03
- name: images
- config_name: CoSyn_400k_nutrition
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 1508994014.353
num_examples: 6931
download_size: 1485416825
dataset_size: 1508994014.353
- name: images
- config_name: CoSyn_400k_table
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 7684052957.968
num_examples: 46518
download_size: 7560488723
dataset_size: 7684052957.968
- name: images
- config_name: DoclingMatix
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 960982374796.28
num_examples: 1270910
download_size: 950807646187
dataset_size: 960982374796.28
- name: images
- config_name: LLaVA_Instruct_150K
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 76726976312.25
num_examples: 157710
download_size: 76639461610
dataset_size: 76726976312.25
- name: images
- config_name: SynthChartNet
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 17908210462
num_examples: 500000
download_size: 17714786653
dataset_size: 17908210462
- name: images
- config_name: SynthCodeNet
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 61998944812.125
num_examples: 499983
download_size: 61472605472
dataset_size: 61998944812.125
- name: images
- config_name: SynthFormulaNet
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 2640399650.375
num_examples: 499997
download_size: 2534243196
dataset_size: 2640399650.375
- name: images
- config_name: Unichart
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 18177703609.375
num_examples: 611925
download_size: 16923868243
dataset_size: 18177703609.375
- name: images
- config_name: a_okvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 22759813096.74
num_examples: 54602
download_size: 22740515076
dataset_size: 22759813096.74
- name: images
- config_name: aguvis-stage-1
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 234712403450.264
num_examples: 458957
download_size: 227839724491
dataset_size: 234712403450.264
- name: images
- config_name: ai2d_merged
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 867183847.75
num_examples: 4858
download_size: 860582630
dataset_size: 867183847.75
- name: images
- config_name: alfworldgpt
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 3890916009.875
num_examples: 45073
download_size: 2887255617
dataset_size: 3890916009.875
- name: images
- config_name: allava_laion
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 366924129181.264
num_examples: 468664
download_size: 366513300480
dataset_size: 366924129181.264
- name: images
- config_name: allava_vflan
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 92699946037.528
num_examples: 177078
download_size: 92524279162
dataset_size: 92699946037.528
- name: images
- config_name: aokvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 896746993.93
num_examples: 16539
download_size: 893471601
dataset_size: 896746993.93
- name: images
- config_name: art
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 5141087027.04
num_examples: 5492
download_size: 5140689948
dataset_size: 5141087027.04
- name: images
- config_name: arxivqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 81923895225
num_examples: 100000
download_size: 81811622698
dataset_size: 81923895225
- name: images
- config_name: bentham
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 1450159090.168
num_examples: 10843
download_size: 1449119505
dataset_size: 1450159090.168
- name: images
- config_name: blockdiagramcomputerized
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 28792412
num_examples: 502
download_size: 28553870
dataset_size: 28792412
- name: images
- config_name: blockdiagramhandwritten
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 146651431.23
num_examples: 1029
download_size: 146191172
dataset_size: 146651431.23
- name: images
- config_name: cambrian(filtered)_processed
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 37993953612.448
num_examples: 83123
download_size: 37702528512
dataset_size: 37993953612.448
- name: images
- config_name: captcha
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1108385677.25
num_examples: 113062
download_size: 1093568723
dataset_size: 1108385677.25
- name: images
- config_name: chart2text
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 1129182569.736
num_examples: 26961
download_size: 1108115443
dataset_size: 1129182569.736
- name: images
- config_name: chartqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 815177635.55
num_examples: 18265
download_size: 803910718
dataset_size: 815177635.55
- name: images
- config_name: chinesememe
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 14244173434.512
num_examples: 54212
download_size: 14222203753
dataset_size: 14244173434.512
- name: images
- config_name: chrome_writting
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 80739342.2
num_examples: 8825
download_size: 79343529
dataset_size: 80739342.2
- name: images
- config_name: clevr
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 10557164224
num_examples: 70000
download_size: 10465001066
dataset_size: 10557164224
- name: images
- config_name: clevr_math
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 9394753178
num_examples: 70000
download_size: 9344480504
dataset_size: 9394753178
- name: images
- config_name: clevr_math(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 708071620.6
num_examples: 5280
download_size: 697997425
dataset_size: 708071620.6
- name: images
- config_name: coco_colors
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 55374513710.125
num_examples: 118287
download_size: 55344845137
dataset_size: 55374513710.125
- name: images
- config_name: cocoqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 2402176655.69
num_examples: 46287
download_size: 2394615512
dataset_size: 2402176655.69
- name: images
- config_name: cocotext
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 7930321103.875
num_examples: 16169
download_size: 7928989554
dataset_size: 7930321103.875
- name: images
- config_name: ctw
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 109319015738.75
num_examples: 24290
download_size: 109306604047
dataset_size: 109319015738.75
- name: images
- config_name: datik
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 3646550812.875
num_examples: 220537
download_size: 3482030545
dataset_size: 3646550812.875
- name: images
- config_name: datikz
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 642401206.67
num_examples: 47441
download_size: 591381396
dataset_size: 642401206.67
- name: images
- config_name: densefusion_1m
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 146400262606.897
num_examples: 1058751
download_size: 144502318476
dataset_size: 146400262606.897
- name: images
- config_name: diagram_image_to_text
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 18704652
num_examples: 300
download_size: 18534456
dataset_size: 18704652
- name: images
- config_name: docvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 12018085125.664
num_examples: 10189
download_size: 12007345171
dataset_size: 12018085125.664
- name: images
- config_name: drivelm
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 37226102202.192
num_examples: 4072
download_size: 34029716036
dataset_size: 37226102202.192
- name: images
- config_name: dvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 4581122677
num_examples: 200000
download_size: 4302544626
dataset_size: 4581122677
- name: images
- config_name: est_vqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 18902348521.25
num_examples: 19358
download_size: 18901853752
dataset_size: 18902348521.25
- name: images
- config_name: face_emotion
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 15127430
num_examples: 797
download_size: 14983116
dataset_size: 15127430
- name: images
- config_name: figureqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 2346521984
num_examples: 100000
download_size: 2222862886
dataset_size: 2346521984
- name: images
- config_name: figureqa(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 419233306.69
num_examples: 17587
download_size: 414996519
dataset_size: 419233306.69
- name: images
- config_name: finqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 138086601.5
num_examples: 5276
download_size: 123625992
dataset_size: 138086601.5
- name: images
- config_name: funsd
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 35306471
num_examples: 194
download_size: 35104326
dataset_size: 35306471
- name: images
- config_name: geo170k(align)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 197478416.875
num_examples: 35297
download_size: 161890724
dataset_size: 197478416.875
- name: images
- config_name: geo170k(qa)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 87625237.87
num_examples: 12101
download_size: 52163819
dataset_size: 87625237.87
- name: images
- config_name: geo3k
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 38756856.17
num_examples: 2091
download_size: 37400382
dataset_size: 38756856.17
- name: images
- config_name: geometry3k(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 197867040.88
num_examples: 9724
download_size: 184961625
dataset_size: 197867040.88
- name: images
- config_name: geomverse
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 1183897660.128
num_examples: 9303
download_size: 1062185395
dataset_size: 1183897660.128
- name: images
- config_name: geoqa+(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 94213384.94
num_examples: 17162
download_size: 90953636
dataset_size: 94213384.94
- name: images
- config_name: geos(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 3936551
num_examples: 498
download_size: 1613784
dataset_size: 3936551
- name: images
- config_name: google_landmarks
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 189460937100.184
num_examples: 299993
download_size: 189343235587
dataset_size: 189460937100.184
- name: images
- config_name: groundui
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 6044682942.056
num_examples: 13531
download_size: 6027988163
dataset_size: 6044682942.056
- name: images
- config_name: handwriting_forms
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 168001610
num_examples: 1400
download_size: 164655119
dataset_size: 168001610
- name: images
- config_name: hateful_memes
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 3059106937.5
num_examples: 8500
download_size: 3058138125
dataset_size: 3059106937.5
- name: images
- config_name: hitab
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 163934179
num_examples: 2500
download_size: 160422628
dataset_size: 163934179
- name: images
- config_name: hme100k
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 1547322234.04
num_examples: 74492
download_size: 1538339958
dataset_size: 1547322234.04
- name: images
- config_name: hw_squad
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 21637654637.632
num_examples: 20457
download_size: 21633468499
dataset_size: 21637654637.632
- name: images
- config_name: iam
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 1138239910.217
num_examples: 5663
download_size: 1134974960
dataset_size: 1138239910.217
- name: images
- config_name: iconqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 330918682.09
num_examples: 27307
download_size: 326819099
dataset_size: 330918682.09
- name: images
- config_name: iconqa(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 209363820.43
num_examples: 22589
download_size: 204676537
dataset_size: 209363820.43
- name: images
- config_name: idk
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 5197458667.01
num_examples: 11123
download_size: 5194521196
dataset_size: 5197458667.01
- name: images
- config_name: iiit5k
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 21788858.44
num_examples: 1990
download_size: 21513252
dataset_size: 21788858.44
- name: images
- config_name: image_textualization(filtered)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 39882386250.375
num_examples: 99573
download_size: 39829746385
dataset_size: 39882386250.375
- name: images
- config_name: imgur5k
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 12591193342.434
num_examples: 5934
download_size: 12590763433
dataset_size: 12591193342.434
- name: images
- config_name: indoor_qa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 797864863.25
num_examples: 3350
download_size: 797431780
dataset_size: 797864863.25
- name: images
- config_name: infographic(gpt4v)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 2014741558.032
num_examples: 1982
download_size: 2011159744
dataset_size: 2014741558.032
- name: images
- config_name: infographic_vqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 4467479648.894
num_examples: 4394
download_size: 4465512444
dataset_size: 4467479648.894
- name: images
- config_name: infographic_vqa_llava_format
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 1765450951.75
num_examples: 2113
download_size: 1764585485
dataset_size: 1765450951.75
- name: images
- config_name: intergps
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 25159455
num_examples: 1280
download_size: 24899065
dataset_size: 25159455
- name: images
- config_name: invoices_receipts
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1925658845.375
num_examples: 3013
download_size: 1923244863
dataset_size: 1925658845.375
- name: images
- config_name: k12_printing
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 4587776492.32
num_examples: 256636
download_size: 4546453791
dataset_size: 4587776492.32
- name: images
- config_name: laion_gpt4v
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 3021991360.375
num_examples: 9301
download_size: 3017230039
dataset_size: 3021991360.375
- name: images
- config_name: latex_handwritten
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 12665387206.408
num_examples: 39583
download_size: 12655091924
dataset_size: 12665387206.408
- name: images
- config_name: latexformulas
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 5604066568.5
num_examples: 552340
download_size: 5525103231
dataset_size: 5604066568.5
- name: images
- config_name: llavar_gpt4_20k
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 4235159184.04
num_examples: 19790
download_size: 4229077598
dataset_size: 4235159184.04
- name: images
- config_name: lnqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 266234680687.28
num_examples: 302780
download_size: 266088073857
dataset_size: 266234680687.28
- name: images
- config_name: localized_narratives
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 21346019807.448
num_examples: 199998
download_size: 21291848742
dataset_size: 21346019807.448
- name: images
- config_name: lrv_chart
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 86444276.008
num_examples: 1776
download_size: 85369432
dataset_size: 86444276.008
- name: images
- config_name: lrv_normal(filtered)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 2985153010.43
num_examples: 10489
download_size: 2967270530
dataset_size: 2985153010.43
- name: images
- config_name: lvis_instruct4v
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 107372123408.125
num_examples: 222711
download_size: 107199700503
dataset_size: 107372123408.125
- name: images
- config_name: mapqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 3371567797.875
num_examples: 37417
download_size: 3308958271
dataset_size: 3371567797.875
- name: images
- config_name: mapqa(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 351524458.75
num_examples: 5225
download_size: 345656246
dataset_size: 351524458.75
- name: images
- config_name: maptext
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1504185688
num_examples: 200
download_size: 1504165598
dataset_size: 1504185688
- name: images
- config_name: mathwriting-google
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 12299849132
num_examples: 300000
download_size: 12219456415
dataset_size: 12299849132
- name: images
- config_name: mavis_math_metagen
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 3975734405.048
num_examples: 87348
download_size: 3266775104
dataset_size: 3975734405.048
- name: images
- config_name: mavis_math_rule_geo
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 20031463665.136
num_examples: 99986
download_size: 19769419782
dataset_size: 20031463665.136
- name: images
- config_name: memotion
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 2530734349.206
num_examples: 6991
download_size: 2528737208
dataset_size: 2530734349.206
- name: images
- config_name: mimic_cgd
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 13184046225.25
num_examples: 70939
download_size: 13149862823
dataset_size: 13184046225.25
- name: images
- config_name: mmc_instruct
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 16504670029.128
num_examples: 168178
download_size: 16230725185
dataset_size: 16504670029.128
- name: images
- config_name: mmevol
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 25742246427.065
num_examples: 160215
download_size: 25480716864
dataset_size: 25742246427.065
- name: images
- config_name: mmra
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1289479228.248
num_examples: 1024
download_size: 1249496994
dataset_size: 1289479228.248
- name: images
- config_name: mmsoc_memotion
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 2531497426.206
num_examples: 6991
download_size: 2529088456
dataset_size: 2531497426.206
- name: images
- config_name: multihiertt
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 1378944031.237
num_examples: 7619
download_size: 1362595573
dataset_size: 1378944031.237
- name: images
- config_name: nlvr2
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 23552929006.152
num_examples: 50426
download_size: 23481437598
dataset_size: 23552929006.152
- name: images
- config_name: objects365_qa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 202609182505.89
num_examples: 1665847
download_size: 199554203410
dataset_size: 202609182505.89
- name: images
- config_name: ocrvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 6148678275.896
num_examples: 165746
download_size: 6057032047
dataset_size: 6148678275.896
- name: images
- config_name: olmOCR-mix-0225-books
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 6633513593.78
num_examples: 15194
download_size: 6618802397
dataset_size: 6633513593.78
- name: images
- config_name: olmOCR-mix-0225-documents
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 97945411922.46
num_examples: 228858
download_size: 97308921712
dataset_size: 97945411922.46
- name: images
- config_name: oodvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 4694217369.688
num_examples: 8488
download_size: 4653237790
dataset_size: 4694217369.688
- name: images
- config_name: orand_car_a
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 23695905.13
num_examples: 1999
download_size: 23351148
dataset_size: 23695905.13
- name: images
- config_name: pathvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 18649772240
num_examples: 32632
download_size: 18155570098
dataset_size: 18649772240
- name: images
- config_name: pdfvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1663168578.91
num_examples: 8593
download_size: 1645451234
dataset_size: 1663168578.91
- name: images
- config_name: plotqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 8939469643.25
num_examples: 157070
download_size: 5345605223
dataset_size: 8939469643.25
- name: images
- config_name: pmc_vqa(mathv360k)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 3445692373.648
num_examples: 35948
download_size: 3437305247
dataset_size: 3445692373.648
- name: images
- config_name: raven
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 1734017137
num_examples: 42000
download_size: 1721095694
dataset_size: 1734017137
- name: images
- config_name: rendered_text
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 11087697572
num_examples: 10000
download_size: 11087197372
dataset_size: 11087697572
- name: images
- config_name: robut_sqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 691188383.864
num_examples: 8514
download_size: 684223334
dataset_size: 691188383.864
- name: images
- config_name: robut_wikisql
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 6319659576.464
num_examples: 74989
download_size: 6292705239
dataset_size: 6319659576.464
- name: images
- config_name: robut_wtq
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 4150727243.896
num_examples: 38246
download_size: 4125713020
dataset_size: 4150727243.896
- name: images
- config_name: scienceqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 287033977.12
num_examples: 4976
download_size: 283309644
dataset_size: 287033977.12
- name: images
- config_name: scienceqa(nona_context)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 2014832143.96
num_examples: 19208
download_size: 1968554064
dataset_size: 2014832143.96
- name: images
- config_name: screen2words
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 1693100006.1
num_examples: 15730
download_size: 1345772929
dataset_size: 1693100006.1
- name: images
- config_name: screenqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 44877746311.875
num_examples: 80761
download_size: 44817901938
dataset_size: 44877746311.875
- name: images
- config_name: sharegpt4o
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 39874535436.384
num_examples: 57284
download_size: 39791929512
dataset_size: 39874535436.384
- name: images
- config_name: sharegpt4v(coco)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 20028717347.875
num_examples: 50017
download_size: 20005211134
dataset_size: 20028717347.875
- name: images
- config_name: sharegpt4v(knowledge)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64
splits: - name: train
num_bytes: 2405546343.928
num_examples: 1988
download_size: 2404816763
dataset_size: 2405546343.928
- name: images
- config_name: sharegpt4v(llava)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 5644424724.75
num_examples: 29986
download_size: 5627968195
dataset_size: 5644424724.75
- name: images
- config_name: sharegpt4v(sam)
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 31591491436.24
num_examples: 8990
download_size: 31588799545
dataset_size: 31591491436.24
- name: images
- config_name: sketchyvqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 462161568
num_examples: 8000
download_size: 454872096
dataset_size: 462161568
- name: images
- config_name: slidevqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 4252726221.347
num_examples: 1919
download_size: 2508659044
dataset_size: 4252726221.347
- name: images
- config_name: spark
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 1062710726.48
num_examples: 3904
download_size: 1061887418
dataset_size: 1062710726.48
- name: images
- config_name: spatialsense
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 3539733377.8
num_examples: 10440
download_size: 3537019555
dataset_size: 3539733377.8
- name: images
- config_name: spot_the_diff
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64
splits: - name: train
num_bytes: 1656404738.5
num_examples: 8566
download_size: 1590994273
dataset_size: 1656404738.5
- name: images
- config_name: sroie
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 382776987.92
num_examples: 33616
download_size: 377976339
dataset_size: 382776987.92
- name: images
- config_name: st_vqa
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64
splits: - name: train
num_bytes: 878506863.672
num_examples: 17247
download_size: 876025784
dataset_size: 878506863.672
- name: images
- config_name: sujet_finance
features:- name: images
list: image - name: texts
list:- name: user
dtype: string - name: assistant
dtype: string
- name: user
- name: source
dtype: string - name: formatting_ratings
list: int64 - name: formatting_min
dtype: int64 - name: visual_dependency_ratings
list: int64 - name: visual_dependency_min
dtype: int64 - name: image_correspondence_ratings
list: int64 - name: image_correspondence_min
dtype: int64 - name: relevance_ratings
list: int64 - name: relevance_min
dtype: int64
splits: - name: train
num_bytes: 4870775458.875
num_examples: 9801
download_size: 4859136094
dataset_size: 4870775458.875
- name: images
[Content truncated...]
Social Proof
AI Summary: Based on Hugging Face metadata. Not a recommendation.
đĄī¸ Dataset Transparency Report
Verified data manifest for traceability and transparency.
đ Identity & Source
- id
- hf-dataset--huggingfacem4--finevision
- source
- huggingface
- author
- HuggingFaceM4
- tags
- size_categories:10m
format:parquetmodality:imagemodality:textlibrary:datasetslibrary:dasklibrary:mlcroissantlibrary:polarsarxiv:2510.17269region:us
âī¸ Technical Specs
- architecture
- null
- params billions
- null
- context length
- null
đ Engagement & Metrics
- likes
- 471
- downloads
- 128,953
Free2AITools Constitutional Data Pipeline: Curated disclosure mode active. (V15.x Standard)