|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: image_id |
|
|
dtype: string |
|
|
- name: label |
|
|
dtype: int32 |
|
|
- name: clip_model |
|
|
dtype: string |
|
|
- name: clip_features |
|
|
list: float32 |
|
|
- name: vector_dim |
|
|
dtype: int32 |
|
|
- name: timestamp |
|
|
dtype: timestamp[ns] |
|
|
splits: |
|
|
- name: clip_vit_b32_train |
|
|
num_bytes: 2723761042 |
|
|
num_examples: 1281167 |
|
|
- name: clip_vit_laion_b32_train |
|
|
num_bytes: 2789100559 |
|
|
num_examples: 1281167 |
|
|
- name: clip_vit_laion_b32_validation |
|
|
num_bytes: 108850000 |
|
|
num_examples: 50000 |
|
|
- name: clip_vit_b16_train |
|
|
num_bytes: 2777570056 |
|
|
num_examples: 1281167 |
|
|
- name: clip_vit_b16_validation |
|
|
num_bytes: 108400000 |
|
|
num_examples: 50000 |
|
|
- name: clip_vit_l14_train |
|
|
num_bytes: 4090766231 |
|
|
num_examples: 1281167 |
|
|
- name: clip_vit_l14_validation |
|
|
num_bytes: 159650000 |
|
|
num_examples: 50000 |
|
|
- name: clip_vit_laion_bigg14_train |
|
|
num_bytes: 6728689084 |
|
|
num_examples: 1281167 |
|
|
- name: clip_vit_laion_bigg14_validation |
|
|
num_bytes: 262600000 |
|
|
num_examples: 50000 |
|
|
- name: clip_vit_b32_validation |
|
|
num_bytes: 108400000 |
|
|
num_examples: 50000 |
|
|
- name: clip_vit_b32_test |
|
|
num_bytes: 216800000 |
|
|
num_examples: 100000 |
|
|
- name: clip_vit_b16_test |
|
|
num_bytes: 216800000 |
|
|
num_examples: 100000 |
|
|
- name: clip_vit_laion_b32_test |
|
|
num_bytes: 217700000 |
|
|
num_examples: 100000 |
|
|
- name: clip_vit_l14_test |
|
|
num_bytes: 319300000 |
|
|
num_examples: 100000 |
|
|
- name: clip_vit_laion_h14_test |
|
|
num_bytes: 422500000 |
|
|
num_examples: 100000 |
|
|
download_size: 25438949728 |
|
|
dataset_size: 21250886972 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: clip_vit_b32_train |
|
|
path: data/clip_vit_b32_train-* |
|
|
- split: clip_vit_b32_validation |
|
|
path: data/clip_vit_b32_validation-* |
|
|
- split: clip_vit_laion_b32_train |
|
|
path: data/clip_vit_laion_b32_train-* |
|
|
- split: clip_vit_laion_b32_validation |
|
|
path: data/clip_vit_laion_b32_validation-* |
|
|
- split: clip_vit_b16_train |
|
|
path: data/clip_vit_b16_train-* |
|
|
- split: clip_vit_b16_validation |
|
|
path: data/clip_vit_b16_validation-* |
|
|
- split: clip_vit_l14_train |
|
|
path: data/clip_vit_l14_train-* |
|
|
- split: clip_vit_l14_validation |
|
|
path: data/clip_vit_l14_validation-* |
|
|
- split: clip_vit_laion_bigg14_train |
|
|
path: data/clip_vit_laion_bigg14_train-* |
|
|
- split: clip_vit_laion_bigg14_validation |
|
|
path: data/clip_vit_laion_bigg14_validation-* |
|
|
- split: clip_vit_b32_test |
|
|
path: data/clip_vit_b32_test-* |
|
|
- split: clip_vit_b16_test |
|
|
path: data/clip_vit_b16_test-* |
|
|
- split: clip_vit_laion_b32_test |
|
|
path: data/clip_vit_laion_b32_test-* |
|
|
- split: clip_vit_l14_test |
|
|
path: data/clip_vit_l14_test-* |
|
|
- split: clip_vit_laion_h14_test |
|
|
path: data/clip_vit_laion_h14_test-* |
|
|
task_categories: |
|
|
- feature-extraction |
|
|
- image-feature-extraction |
|
|
license: mit |
|
|
tags: |
|
|
- features |
|
|
- image_features |
|
|
- extracted_features |
|
|
- precomputed_features |
|
|
- imagenet |
|
|
- imagenet_features |
|
|
- clip_vit |
|
|
- variants |
|
|
size_categories: |
|
|
- 1M<n<10M |
|
|
--- |
|
|
|
|
|
# Update: 10/2/2025 |
|
|
|
|
|
Claude said that I'm not being careful enough with my database curation after grilling me for 20 minutes, so I included the preparer script as well. |
|
|
|
|
|
Claude Sonnet 4.5 is kind of a chad. |
|
|
|
|
|
# Update; 9/26/2025 |
|
|
|
|
|
Having to download this whole repo is annoying, so I'm making sure the splits are named train/val/test (if they exist) and the named subset is the clip name. |
|
|
|
|
|
|
|
|
# Older non-dated updates |
|
|
|
|
|
Everything extracted with torch configured as deterministic; using seed 42 on an a100 using colab; so if it has variances from expectation it's on cuda. |
|
|
|
|
|
It's a little quirky; |
|
|
* Most of the splits have train, test, val. Many do not. |
|
|
* Most of the splits have a proper "image_id" md5 id for verification. |
|
|
|
|
|
The prompts used were direct literal prompts for the classification name; |
|
|
|
|
|
No use of "a photo of" or any such invariance; just the classification text. |
|
|
|
|
|
This is a series of clip-vit extracted feature maps from a 256x256 cropped and resized imagenet variant hosted here on huggingface. |
|
|
|
|
|
I ran the processor 224x224 and then extracted features from the entire dataset batch-sequentially while simultaneously capturing the necessary classifiers and classifications associated with the images for downstream testing and assessment. |
|
|
|
|
|
Academic and research purpose use only. |
|
|
|
|
|
clip-vit-large-patch14 variations do exist in the splits. |
|
|
|
|
|
clip-vit-bigG is the 1280 dim variation and it does exist; it took quite a while to extract - and it is in fact missing it's test split. Sorry about that. |
|
|
|
|
|
There are many variants of clip-vit-base from many variant forms. Each of them extracted using the same process as the others. |
|
|
|