1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
|
- """Main"""
- import torch
- import torch.nn as nn
- import torch.optim as optim
- from torch.utils.data import DataLoader
- from datetime import datetime
- import matplotlib
- #matplotlib.use('agg')
- from matplotlib import pyplot as plt
- plt.ion()
- import os
- import sys
- from dataset.dataset_class import PreprocessDataset
- from dataset.video_extraction_conversion import *
- from loss.loss_discriminator import *
- from loss.loss_generator import *
- from network.blocks import *
- from network.model import *
- from tqdm import tqdm
- from params.params import K, path_to_chkpt, path_to_backup, path_to_Wi, batch_size, path_to_preprocess, frame_shape
- """Create dataset and net"""
- display_training = False
- device = torch.device("cuda:0")
- cpu = torch.device("cpu")
- dataset = PreprocessDataset(K=K, path_to_preprocess=path_to_preprocess, path_to_Wi=path_to_Wi)
- dataLoader = DataLoader(dataset, batch_size=32, shuffle=True, num_workers=16, pin_memory=True)
- E = nn.DataParallel(Embedder(frame_shape).to(device))
- """Training init"""
- epochCurrent = epoch = i_batch = 0
- lossesG = []
- lossesD = []
- i_batch_current = 0
- num_epochs = 1
- #initiate checkpoint if inexistant
- if not os.path.isfile(path_to_chkpt):
- print('Error loading model: file non-existant')
- sys.exit()
- """Loading from past checkpoint"""
- checkpoint = torch.load(path_to_chkpt, map_location=cpu)
- E.module.load_state_dict(checkpoint['E_state_dict'])
- num_vid = checkpoint['num_vid']
- E.train(False)
- #init W_i
- print('Initializing Discriminator weights')
- if not os.path.isdir(path_to_Wi):
- os.mkdir(path_to_Wi)
- for i in tqdm(range(num_vid)):
- if not os.path.isfile(path_to_Wi+'/W_'+str(i)+'/W_'+str(i)+'.tar'):
- w_i = torch.rand(512, 1)
- os.mkdir(path_to_Wi+'/W_'+str(i))
- torch.save({'W_i': w_i}, path_to_Wi+'/W_'+str(i)+'/W_'+str(i)+'.tar')
- """Training"""
- batch_start = datetime.now()
- pbar = tqdm(dataLoader, leave=True, initial=0)
- if not display_training:
- matplotlib.use('agg')
- with torch.no_grad():
- for epoch in range(num_epochs):
- if epoch > epochCurrent:
- i_batch_current = 0
- pbar = tqdm(dataLoader, leave=True, initial=0)
- pbar.set_postfix(epoch=epoch)
- for i_batch, (f_lm, x, g_y, i, W_i) in enumerate(pbar, start=0):
-
- f_lm = f_lm.to(device)
-
- #zero the parameter gradients
-
- #forward
- # Calculate average encoding vector for video
- f_lm_compact = f_lm.view(-1, f_lm.shape[-4], f_lm.shape[-3], f_lm.shape[-2], f_lm.shape[-1]) #BxK,2,3,224,224
-
- e_vectors = E(f_lm_compact[:,0,:,:,:], f_lm_compact[:,1,:,:,:]) #BxK,512,1
- e_vectors = e_vectors.view(-1, f_lm.shape[1], 512, 1) #B,K,512,1
- e_hat = e_vectors.mean(dim=1)
- for enum, idx in enumerate(i):
- torch.save({'W_i': e_hat[enum,:].unsqueeze(0)}, path_to_Wi+'/W_'+str(idx.item())+'/W_'+str(idx.item())+'.tar')
|