in pytouch/datasets/slip_sequence.py [0:0]
def __getitem__(self, index):
sequence_path = os.path.join(self.root, self.videos[index])
frame_indices = self.default_frame_indices
if self.temporal_transform is not None:
frame_indices = self.temporal_transform(frame_indices)
# list of frames
clip = self.loader(sequence_path, frame_indices)
if self.spatial_transform is not None:
self.spatial_transform.randomize_parameters()
clip = [self.spatial_transform(img) for img in clip]
clip = torch.stack(clip, 0).permute(1, 0, 2, 3)
target = self.annotations[index]
if self.target_transform is not None:
target = self.target_transform
return clip, target