-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathdataloader.py
More file actions
62 lines (45 loc) · 2.41 KB
/
dataloader.py
File metadata and controls
62 lines (45 loc) · 2.41 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
from torchvision import transforms
import torchvision
import torch
from torch.utils.data import Dataset
from torchvision.transforms import InterpolationMode
from tqdm import tqdm
class CustomDataset(Dataset):
def __init__(self, root_dir, isTrain):
if isTrain:
data_augmentation = transforms.Compose([transforms.Resize((600, 600), InterpolationMode.BILINEAR),
transforms.RandomCrop((448, 448)),
transforms.RandomHorizontalFlip(),
transforms.ToTensor(),
transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])])
else:
data_augmentation = transforms.Compose([transforms.Resize((600, 600), InterpolationMode.BILINEAR),
transforms.CenterCrop((448, 448)),
transforms.ToTensor(),
transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])])
dataset = torchvision.datasets.ImageFolder(root_dir, transform=data_augmentation)
self.train_dataset = []
self.val_dataset = []
# 일단 순서대로 저장이 돼 있을텐데 하나는 txt파일에 따라 순서대로 traindatset testdataset에 사진과 index를 넣어줄 것이다.
line = read_txt('./CUB_200_2011/train_test_split.txt')
# testing short version
#line = line[0:int(len(line)/3)]
print('[Training_data]')
self.train_dataset = [dataset[i] for i in tqdm(range(len(line))) if line[i][-1] == '1']
print()
print('[Validating_data]')
self.val_dataset = [dataset[i] for i in tqdm(range(len(line))) if line[i][-1] == '0']
def __len__(self):
return len(self.train_dataset), len(self.val_dataset)
def __getitem__(self, idx):
return self.train_dataset[idx], self.val_dataset[idx]
def read_txt(root_dir):
f = open(root_dir, mode='r')
line = f.read().split('\n')
if line[-1] == '' or ' ':
del line[-1]
line = sorted(line, key=lambda x: int(x.split(' ')[0]))
return line
def Dataloader(dataset, batch_size):
dataloader = torch.utils.data.DataLoader(dataset, batch_size=batch_size, shuffle=True, num_workers=0)
return dataloader