{"id":296,"slug":"mvp-lab--llava-onevision-1.5-instruct-data","name":"LLaVA-OneVision-1.5-Instruct-Data","author":"mvp-lab","description":"\n\t\n\t\t\n\t\tLLaVA-OneVision-1.5 Instruction Data\n\t\n\nPaper | Code\n\n\t\n\t\t\n\t\t📌 Introduction\n\t\n\nThis dataset, LLaVA-OneVision-1.5-Instruct, was collected and integrated during the development of LLaVA-OneVision-1.5. LLaVA-OneVision-1.5 is a novel family of Large Multimodal Models (LMMs) that achieve state-of-the-art performance with significantly reduced computational and financial costs. This meticulously curated 22M instruction dataset (LLaVA-OneVision-1.5-Instruct) is part of a comprehensive and… See the full description on the dataset page: https://huggingface.co/datasets/mvp-lab/LLaVA-OneVision-1.5-Instruct-Data.","tags":"[\"Task_categories:image-Text-To-Text\",\"Language:en\",\"Size_categories:10M<n<100M\",\"Modality:image\",\"Modality:text\",\"Multimodal\"]","license":null,"framework":null,"parameters":null,"downloads":89804,"likes":71,"verified":0,"created_at":"2026-04-20 21:02:07","updated_at":"2026-05-08 06:10:30","source_url":"https://huggingface.co/datasets/mvp-lab/LLaVA-OneVision-1.5-Instruct-Data","source_platform":"huggingface","hf_repo_id":"mvp-lab/LLaVA-OneVision-1.5-Instruct-Data","ollama_name":"","category":"dataset","latest_version":"v1.0.0","version_count":1,"signature_count":1,"risk_level":null,"risk_score":null,"versions":[{"id":295,"model_id":296,"version":"v1.0.0","manifest_hash":"509f0a9001130bad25120388cfd6efe00b1f9f827875897c9e6d4f9cf827f93a","file_count":0,"total_size":0,"r2_manifest_key":"manifests/datasets/mvp-lab--llava-onevision-1.5-instruct-data/v1.0.0.json","created_at":"2026-04-20 21:02:07"}],"files":[],"signatures":[{"id":752,"version_id":295,"signer_did":"did:quantamrkt:registry:shield-v1","algorithm":"ML-DSA-65","signature_hex":"886968426272988fdebeee092827cdc6be7ea0c312a7a1a6714e4a91ab140ff2","attestation_type":"registry","signed_at":"2026-04-20 21:02:07"}],"hndl":null}