mirror of
https://github.com/AntoineHX/smart_augmentation.git
synced 2025-05-04 12:10:45 +02:00
Rangement
This commit is contained in:
parent
ca3367d19f
commit
4166922c34
453 changed files with 9797 additions and 7 deletions
0
higher/smart_aug/benchmark.py
Normal file → Executable file
0
higher/smart_aug/benchmark.py
Normal file → Executable file
0
higher/smart_aug/higher_patch.py
Normal file → Executable file
0
higher/smart_aug/higher_patch.py
Normal file → Executable file
0
higher/smart_aug/old/compare_TF.py
Normal file → Executable file
0
higher/smart_aug/old/compare_TF.py
Normal file → Executable file
0
higher/smart_aug/old/dataug_old.py
Normal file → Executable file
0
higher/smart_aug/old/dataug_old.py
Normal file → Executable file
0
higher/smart_aug/old/higher_repro.py
Normal file → Executable file
0
higher/smart_aug/old/higher_repro.py
Normal file → Executable file
0
higher/smart_aug/old/model_old.py
Normal file → Executable file
0
higher/smart_aug/old/model_old.py
Normal file → Executable file
0
higher/smart_aug/old/train_utils_old.py
Normal file → Executable file
0
higher/smart_aug/old/train_utils_old.py
Normal file → Executable file
0
higher/smart_aug/old/utils_old.py
Normal file → Executable file
0
higher/smart_aug/old/utils_old.py
Normal file → Executable file
0
higher/smart_aug/smart_aug_example.py
Normal file → Executable file
0
higher/smart_aug/smart_aug_example.py
Normal file → Executable file
|
@ -8,7 +8,7 @@ from dataug import *
|
|||
from train_utils import *
|
||||
from transformations import TF_loader
|
||||
|
||||
postfix='-metaScheduler'
|
||||
postfix='-MetaScheduler2'
|
||||
TF_loader=TF_loader()
|
||||
|
||||
device = torch.device('cuda') #Select device to use
|
||||
|
@ -34,13 +34,13 @@ if __name__ == "__main__":
|
|||
}
|
||||
#Parameters
|
||||
n_inner_iter = 1
|
||||
epochs = 200
|
||||
epochs = 100
|
||||
dataug_epoch_start=0
|
||||
Nb_TF_seq=3
|
||||
optim_param={
|
||||
'Meta':{
|
||||
'optim':'Adam',
|
||||
'lr':1e-4, #1e-2
|
||||
'lr':1e-3, #1e-2
|
||||
'epoch_start': 2, #0 / 2 (Resnet?)
|
||||
'reg_factor': 0.001,
|
||||
'scheduler': 'multiStep', #None, 'multiStep'
|
||||
|
|
|
@ -295,8 +295,8 @@ def run_dist_dataugV3(model, opt_param, epochs=1, inner_it=1, dataug_epoch_start
|
|||
meta_scheduler=None
|
||||
if opt_param['Meta']['scheduler']=='multiStep':
|
||||
meta_scheduler=torch.optim.lr_scheduler.MultiStepLR(meta_opt,
|
||||
milestones=[int(epochs/3), int(epochs*2/3), int(epochs*2.7/3)],
|
||||
gamma=10)
|
||||
milestones=[int(epochs/3), int(epochs*2/3)]#, int(epochs*2.7/3)],
|
||||
gamma=3.16)#10)
|
||||
elif opt_param['Meta']['scheduler'] is not None:
|
||||
raise ValueError("Lr scheduler unknown : %s"%opt_param['Meta']['scheduler'])
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue