-
Notifications
You must be signed in to change notification settings - Fork 15
Expand file tree
/
Copy pathtrain.py
More file actions
94 lines (79 loc) · 3 KB
/
train.py
File metadata and controls
94 lines (79 loc) · 3 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
import argparse
from torch.utils.data import DataLoader
from lightning.pytorch import Trainer
from lightning.pytorch.callbacks import ModelCheckpoint
from avdeepfake1m.loader import AVDeepfake1mPlusPlusImages
from xception import Xception
from utils import LrLogger, EarlyStoppingLR
parser = argparse.ArgumentParser(description="Classification model training")
parser.add_argument("--data_root", type=str)
parser.add_argument("--batch_size", type=int, default=128)
parser.add_argument("--model", type=str, choices=["xception", "meso4", "meso_inception4"])
parser.add_argument("--gpus", type=int, default=1)
parser.add_argument("--precision", default=32)
parser.add_argument("--num_train", type=int, default=None)
parser.add_argument("--num_val", type=int, default=2000)
parser.add_argument("--max_epochs", type=int, default=500)
parser.add_argument("--resume", type=str, default=None)
args = parser.parse_args()
if __name__ == "__main__":
# You can fix the random seed if you want reproducible subsets each epoch:
# torch.manual_seed(42)
# random.seed(42)
learning_rate = 1e-4
gpus = args.gpus
total_batch_size = args.batch_size * gpus
learning_rate = learning_rate * total_batch_size / 4
# Setup model
if args.model == "xception":
model = Xception(learning_rate, distributed=gpus > 1)
else:
raise ValueError(f"Unknown model: {args.model}")
train_dataset = AVDeepfake1mPlusPlusImages(
subset="train",
data_root=args.data_root,
take_num=args.num_train,
use_video_label=True # For video-level label access, set True
)
# For validation, you can still do the normal dataset
val_dataset = AVDeepfake1mPlusPlusImages(
subset="val",
data_root=args.data_root,
take_num=args.num_val,
use_video_label=True
)
# Parse precision properly
try:
precision = int(args.precision)
except ValueError:
precision = args.precision
monitor = "val_loss"
trainer = Trainer(
log_every_n_steps=50,
precision=precision,
max_epochs=args.max_epochs,
callbacks=[
ModelCheckpoint(
dirpath=f"./ckpt/{args.model}",
save_last=True,
filename=args.model + "-{epoch}-{val_loss:.3f}",
monitor=monitor,
mode="min"
),
LrLogger(),
EarlyStoppingLR(lr_threshold=1e-7)
],
enable_checkpointing=True,
benchmark=True,
accelerator="gpu",
devices=args.gpus,
strategy="ddp" if args.gpus > 1 else "auto",
# ckpt_path=args.resume,
# If you're on an older version of Lightning, you may need `strategy='ddp'` just the same, but this is typical.
)
trainer.fit(
model,
train_dataloaders=DataLoader(train_dataset, batch_size=args.batch_size, num_workers=0),
val_dataloaders=DataLoader(val_dataset, batch_size=args.batch_size, num_workers=0),
ckpt_path=args.resume,
)