|
import os, io, csv, math, random |
|
import numpy as np |
|
from einops import rearrange |
|
|
|
import torch |
|
from decord import VideoReader |
|
import cv2 |
|
from scipy.ndimage import distance_transform_edt |
|
import torchvision.transforms as transforms |
|
from torch.utils.data.dataset import Dataset |
|
|
|
|
|
from PIL import Image |
|
def pil_image_to_numpy(image, is_maks = False, index = 1): |
|
"""Convert a PIL image to a NumPy array.""" |
|
|
|
if is_maks: |
|
|
|
image = image.resize((256, 256)) |
|
|
|
|
|
return np.array(image) |
|
else: |
|
if image.mode != 'RGB': |
|
image = image.convert('RGB') |
|
image = image.resize((256, 256)) |
|
return np.array(image) |
|
|
|
def numpy_to_pt(images: np.ndarray, is_mask=False) -> torch.FloatTensor: |
|
"""Convert a NumPy image to a PyTorch tensor.""" |
|
if images.ndim == 3: |
|
images = images[..., None] |
|
images = torch.from_numpy(images.transpose(0, 3, 1, 2)) |
|
if is_mask: |
|
return images.float() |
|
else: |
|
return images.float() / 255 |
|
|
|
|
|
class WebVid10M(Dataset): |
|
def __init__( |
|
self,video_folder,ann_folder,motion_folder, |
|
sample_size=256, sample_stride=4, sample_n_frames=14, |
|
): |
|
|
|
self.dataset = [i for i in os.listdir(video_folder)] |
|
|
|
self.length = len(self.dataset) |
|
print(f"data scale: {self.length}") |
|
random.shuffle(self.dataset) |
|
self.video_folder = video_folder |
|
self.sample_stride = sample_stride |
|
self.sample_n_frames = sample_n_frames |
|
self.ann_folder = ann_folder |
|
self.heatmap = self.gen_gaussian_heatmap() |
|
self.motion_values_folder=motion_folder |
|
self.sample_size = sample_size |
|
print("length",len(self.dataset)) |
|
sample_size = tuple(sample_size) if not isinstance(sample_size, int) else (sample_size, sample_size) |
|
|
|
print("sample size",sample_size) |
|
self.pixel_transforms = transforms.Compose([ |
|
|
|
transforms.Resize(sample_size), |
|
|
|
transforms.Normalize(mean=[0.5, 0.5, 0.5], std=[0.5, 0.5, 0.5], inplace=True), |
|
]) |
|
|
|
|
|
|
|
def center_crop(self,img): |
|
h, w = img.shape[-2:] |
|
min_dim = min(h, w) |
|
top = (h - min_dim) // 2 |
|
left = (w - min_dim) // 2 |
|
return img[..., top:top+min_dim, left:left+min_dim] |
|
|
|
def gen_gaussian_heatmap(self,imgSize=200): |
|
circle_img = np.zeros((imgSize, imgSize), np.float32) |
|
circle_mask = cv2.circle(circle_img, (imgSize//2, imgSize//2), imgSize//2, 1, -1) |
|
|
|
|
|
isotropicGrayscaleImage = np.zeros((imgSize, imgSize), np.float32) |
|
|
|
|
|
for i in range(imgSize): |
|
for j in range(imgSize): |
|
isotropicGrayscaleImage[i, j] = 1 / 2 / np.pi / (40 ** 2) * np.exp( |
|
-1 / 2 * ((i - imgSize / 2) ** 2 / (40 ** 2) + (j - imgSize / 2) ** 2 / (40 ** 2))) |
|
|
|
|
|
isotropicGrayscaleImage = isotropicGrayscaleImage * circle_mask |
|
isotropicGrayscaleImage = (isotropicGrayscaleImage / np.max(isotropicGrayscaleImage)).astype(np.float32) |
|
isotropicGrayscaleImage = (isotropicGrayscaleImage / np.max(isotropicGrayscaleImage)*255).astype(np.uint8) |
|
|
|
|
|
return isotropicGrayscaleImage |
|
|
|
def calculate_center_coordinates(self,masks,ids, side=20): |
|
center_coordinates = [] |
|
ids = random.choice(ids[1:]) |
|
for index_mask, mask in enumerate(masks): |
|
new_img = np.zeros((self.sample_size, self.sample_size), np.float32) |
|
|
|
|
|
|
|
|
|
|
|
|
|
for index in [ids]: |
|
mask_array = (np.array(mask)==index)*1 |
|
|
|
|
|
distance_transform = distance_transform_edt(mask_array) |
|
center_coordinate = np.unravel_index(np.argmax(distance_transform), distance_transform.shape) |
|
|
|
y1 = max(center_coordinate[0]-side,0) |
|
y2 = min(center_coordinate[0]+side,self.sample_size-1) |
|
x1 = max(center_coordinate[1]-side,0) |
|
x2 = min(center_coordinate[1]+side,self.sample_size-1) |
|
need_map = cv2.resize(self.heatmap, (x2-x1, y2-y1)) |
|
new_img[y1:y2,x1:x2] = need_map |
|
if index_mask == 0: |
|
new_img = mask_array*255 |
|
new_img = cv2.cvtColor(new_img.astype(np.uint8), cv2.COLOR_GRAY2RGB) |
|
|
|
center_coordinates.append(new_img) |
|
return center_coordinates |
|
|
|
def get_batch(self, idx): |
|
def sort_frames(frame_name): |
|
return int(frame_name.split('.')[0]) |
|
|
|
while True: |
|
videoid = self.dataset[idx] |
|
|
|
|
|
preprocessed_dir = os.path.join(self.video_folder, videoid) |
|
ann_folder = os.path.join(self.ann_folder, videoid) |
|
motion_values_file = os.path.join(self.motion_values_folder, videoid, videoid + "_average_motion.txt") |
|
|
|
if not os.path.exists(ann_folder): |
|
idx = random.randint(0, len(self.dataset) - 1) |
|
continue |
|
|
|
|
|
image_files = sorted(os.listdir(preprocessed_dir), key=sort_frames)[:14] |
|
depth_files = sorted(os.listdir(ann_folder), key=sort_frames)[:14] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
numpy_images = np.array([pil_image_to_numpy(Image.open(os.path.join(preprocessed_dir, img))) for img in image_files]) |
|
pixel_values = numpy_to_pt(numpy_images) |
|
|
|
|
|
mask = Image.open(os.path.join(ann_folder, depth_files[0])).convert('P') |
|
ids = [i for i in np.unique(mask)] |
|
|
|
if len(ids)==1: |
|
idx = random.randint(0, len(self.dataset) - 1) |
|
continue |
|
|
|
numpy_depth_images = np.array([pil_image_to_numpy(Image.open(os.path.join(ann_folder, df)).convert('P'),True,ids) for df in depth_files]) |
|
|
|
heatmap_pixel_values = np.array(self.calculate_center_coordinates(numpy_depth_images,ids)) |
|
|
|
mask_pixel_values = numpy_to_pt(numpy_depth_images,True) |
|
heatmap_pixel_values = numpy_to_pt(heatmap_pixel_values,True) |
|
|
|
|
|
|
|
motion_values = 180 |
|
|
|
|
|
|
|
return pixel_values, mask_pixel_values, motion_values, heatmap_pixel_values |
|
|
|
|
|
|
|
|
|
def __len__(self): |
|
return self.length |
|
|
|
def coordinates_normalize(self,center_coordinates): |
|
first_point = center_coordinates[0] |
|
center_coordinates = [one-first_point for one in center_coordinates] |
|
|
|
return center_coordinates |
|
|
|
def normalize(self, images): |
|
""" |
|
Normalize an image array to [-1,1]. |
|
""" |
|
return 2.0 * images - 1.0 |
|
|
|
def __getitem__(self, idx): |
|
|
|
|
|
|
|
pixel_values, depth_pixel_values,motion_values,heatmap_pixel_values = self.get_batch(idx) |
|
|
|
|
|
|
|
|
|
|
|
pixel_values = self.normalize(pixel_values) |
|
|
|
sample = dict(pixel_values=pixel_values, depth_pixel_values=depth_pixel_values, |
|
motion_values=motion_values,heatmap_pixel_values=heatmap_pixel_values) |
|
return sample |
|
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
from util import save_videos_grid |
|
|
|
dataset = WebVid10M( |
|
video_folder = "/mmu-ocr/weijiawu/MovieDiffusion/svd-temporal-controlnet/data/ref-youtube-vos/train/JPEGImages", |
|
ann_folder = "/mmu-ocr/weijiawu/MovieDiffusion/svd-temporal-controlnet/data/ref-youtube-vos/train/Annotations", |
|
motion_folder = "", |
|
sample_size=256, |
|
sample_stride=1, sample_n_frames=16 |
|
) |
|
|
|
|
|
|
|
dataloader = torch.utils.data.DataLoader(dataset, batch_size=1, num_workers=16,) |
|
for idx, batch in enumerate(dataloader): |
|
images = ((batch["pixel_values"][0].permute(0,2,3,1)+1)/2)*255 |
|
masks = batch["depth_pixel_values"][0].permute(0,2,3,1)*255 |
|
heatmaps = batch["heatmap_pixel_values"][0].permute(0,2,3,1) |
|
|
|
|
|
print(batch["pixel_values"].shape) |
|
|
|
for i in range(images.shape[0]): |
|
image = images[i].numpy().astype(np.uint8) |
|
mask = masks[i].numpy() |
|
heatmap = heatmaps[i].numpy() |
|
|
|
|
|
|
|
|
|
|
|
|
|
print(np.unique(mask)) |
|
cv2.imwrite("./vis/image_{}.jpg".format(i), image) |
|
cv2.imwrite("./vis/mask_{}.jpg".format(i), mask.astype(np.uint8)) |
|
cv2.imwrite("./vis/heatmap_{}.jpg".format(i), heatmap.astype(np.uint8)) |
|
cv2.imwrite("./vis/{}.jpg".format(i), heatmap.astype(np.uint8)*0.5+image*0.5) |
|
|
|
break |