[278d8a]: / reproducibility / embedders / internal_datasets.py

Download this file

58 lines (43 with data), 1.7 kB

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
from torch.utils.data import Dataset
from PIL import Image, ImageFile
ImageFile.LOAD_TRUNCATED_IMAGES = True
Image.MAX_IMAGE_PIXELS = None
class CLIPImageCaptioningDataset(Dataset):
def __init__(self, df, preprocessing):
self.images = df["image"].tolist()
self.caption = df["caption"].tolist()
self.preprocessing = preprocessing
def __len__(self):
return len(self.caption)
def __getitem__(self, idx):
images = self.preprocessing(Image.open(self.images[idx]).convert('RGB')) # preprocess from clip.load
caption = self.caption[idx]
return images, caption
class CLIPCaptioningDataset(Dataset):
def __init__(self, captions):
self.caption = captions
def __len__(self):
return len(self.caption)
def __getitem__(self, idx):
caption = self.caption[idx]
return caption
class CLIPImageDataset(Dataset):
def __init__(self, list_of_images, preprocessing):
self.images = list_of_images
self.preprocessing = preprocessing
def __len__(self):
return len(self.images)
def __getitem__(self, idx):
images = self.preprocessing(Image.open(self.images[idx]).convert('RGB')) # preprocess from clip.load
return images
class CLIPImageLabelDataset(Dataset):
def __init__(self, df, preprocessing):
self.images = df["image"].tolist()
self.label = df["label"].tolist()
self.preprocessing = preprocessing
def __len__(self):
return len(self.images)
def __getitem__(self, idx):
images = self.preprocessing(Image.open(self.images[idx]).convert('RGB')) # preprocess from clip.load
label = self.label[idx]
return images, label