Python torch.optim.Adamax() Examples
The following are 30
code examples of torch.optim.Adamax().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
torch.optim
, or try the search function
.
Example #1
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #2
Source File: model.py From justcopy-backend with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #3
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #4
Source File: make_optimizer.py From MAMS-for-ABSA with Apache License 2.0 | 6 votes |
def make_optimizer(config, model): mode = config['mode'] config = config['aspect_' + mode + '_model'][config['aspect_' + mode + '_model']['type']] lr = config['learning_rate'] weight_decay = config['weight_decay'] opt = { 'sgd': optim.SGD, 'adadelta': optim.Adadelta, 'adam': optim.Adam, 'adamax': optim.Adamax, 'adagrad': optim.Adagrad, 'asgd': optim.ASGD, 'rmsprop': optim.RMSprop, 'adabound': adabound.AdaBound } if 'momentum' in config: optimizer = opt[config['optimizer']](model.parameters(), lr=lr, weight_decay=weight_decay, momentum=config['momentum']) else: optimizer = opt[config['optimizer']](model.parameters(), lr=lr, weight_decay=weight_decay) return optimizer
Example #5
Source File: optimizer.py From XenonPy with BSD 3-Clause "New" or "Revised" License | 6 votes |
def __init__(self, *, lr=0.002, betas=(0.9, 0.999), eps=1e-08, weight_decay=0): """Implements Adamax algorithm (a variant of Adam based on infinity norm). It has been proposed in `Adam: A Method for Stochastic Optimization`__. Arguments: lr (float, optional): learning rate (default: 2e-3) betas (Tuple[float, float], optional): coefficients used for computing running averages of gradient and its square eps (float, optional): term added to the denominator to improve numerical stability (default: 1e-8) weight_decay (float, optional): weight decay (L2 penalty) (default: 0) __ https://arxiv.org/abs/1412.6980 """ super().__init__(optim.Adamax, lr=lr, betas=betas, eps=eps, weight_decay=weight_decay)
Example #6
Source File: optimization.py From SeaRNN-open with MIT License | 6 votes |
def create_optimizer(parameters, opt): lr = opt.learning_rate # default learning rates: # sgd - 0.5, adagrad - 0.01, adadelta - 1, adam - 0.001, adamax - 0.002, asgd - 0.01, rmsprop - 0.01, rprop - 0.01 optim_method = opt.optim_method.casefold() if optim_method == 'sgd': optimizer = optim.SGD(parameters, lr=lr if lr else 0.5, weight_decay=opt.weight_decay) elif optim_method == 'adagrad': optimizer = optim.Adagrad(parameters, lr=lr if lr else 0.01, weight_decay=opt.weight_decay) elif optim_method == 'adadelta': optimizer = optim.Adadelta(parameters, lr=lr if lr else 1, weight_decay=opt.weight_decay) elif optim_method == 'adam': optimizer = optim.Adam(parameters, lr=lr if lr else 0.001, weight_decay=opt.weight_decay) elif optim_method == 'adamax': optimizer = optim.Adamax(parameters, lr=lr if lr else 0.002, weight_decay=opt.weight_decay) elif optim_method == 'asgd': optimizer = optim.ASGD(parameters, lr=lr if lr else 0.01, t0=5000, weight_decay=opt.weight_decay) elif optim_method == 'rmsprop': optimizer = optim.RMSprop(parameters, lr=lr if lr else 0.01, weight_decay=opt.weight_decay) elif optim_method == 'rprop': optimizer = optim.Rprop(parameters, lr=lr if lr else 0.01) else: raise RuntimeError("Invalid optim method: " + opt.optim_method) return optimizer
Example #7
Source File: model.py From RL-based-Graph2Seq-for-NQG with Apache License 2.0 | 6 votes |
def _init_optimizer(self): parameters = [p for p in self.network.parameters() if p.requires_grad] if self.config['use_bert'] and self.config.get('finetune_bert', None): parameters += [p for p in self.config['bert_model'].parameters() if p.requires_grad] if self.config['optimizer'] == 'sgd': self.optimizer = optim.SGD(parameters, self.config['learning_rate'], momentum=self.config['momentum'], weight_decay=self.config['weight_decay']) elif self.config['optimizer'] == 'adam': self.optimizer = optim.Adam(parameters, lr=self.config['learning_rate']) elif self.config['optimizer'] == 'adamax': self.optimizer = optim.Adamax(parameters, lr=self.config['learning_rate']) else: raise RuntimeError('Unsupported optimizer: %s' % self.config['optimizer']) self.scheduler = ReduceLROnPlateau(self.optimizer, mode='max', factor=0.5, \ patience=2, verbose=True)
Example #8
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #9
Source File: test_trainer.py From snorkel with Apache License 2.0 | 6 votes |
def test_optimizer_init(self): trainer = Trainer(**base_config, optimizer="sgd") trainer.fit(model, [dataloaders[0]]) self.assertIsInstance(trainer.optimizer, optim.SGD) trainer = Trainer(**base_config, optimizer="adam") trainer.fit(model, [dataloaders[0]]) self.assertIsInstance(trainer.optimizer, optim.Adam) trainer = Trainer(**base_config, optimizer="adamax") trainer.fit(model, [dataloaders[0]]) self.assertIsInstance(trainer.optimizer, optim.Adamax) with self.assertRaisesRegex(ValueError, "Unrecognized optimizer"): trainer = Trainer(**base_config, optimizer="foo") trainer.fit(model, [dataloaders[0]])
Example #10
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters,lr=2e-3, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #11
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #12
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #13
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #14
Source File: model.py From RCZoo with MIT License | 6 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #15
Source File: SDNetTrainer.py From SDNet with MIT License | 6 votes |
def setup_model(self, vocab_embedding): self.train_loss = AverageMeter() self.network = SDNet(self.opt, vocab_embedding) if self.use_cuda: self.log('Putting model into GPU') self.network.cuda() parameters = [p for p in self.network.parameters() if p.requires_grad] self.optimizer = optim.Adamax(parameters) if 'ADAM2' in self.opt: print('ADAM2') self.optimizer = optim.Adam(parameters, lr = 0.0001) self.updates = 0 self.epoch_start = 0 self.loss_func = F.cross_entropy
Example #16
Source File: model_runner.py From ACMN-Pytorch with MIT License | 5 votes |
def __init__(self, model, model_opt, forward_fn, optimizer, lr_scheduler = None): super(ModelRunner, self).__init__() self.model = model(model_opt) self.optimizer = optimizer self.lr_scheduler = lr_scheduler if self.optimizer == 'adam' : self.optimizer = optim.Adam(self.model.parameters() , \ lr=model_opt.lr , \ betas=(model_opt.beta1, model_opt.beta2), \ weight_decay=model_opt.weight_decay) elif self.optimizer == 'sgd': self.optimizer = optim.SGD(self.model.parameters() , \ lr=model_opt.lr , \ momentum = model_opt.momentum , \ weight_decay=model_opt.weight_decay) elif self.optimizer == 'adamax': self.optimizer = optim.Adamax(self.model.parameters(), lr = model_opt.lr) if lr_scheduler == 'step' : self.lr_scheduler = torch.optim.lr_scheduler.StepLR(self.optimizer, model_opt.lr_step_size, model_opt.lr_gamma) self.gpu = model_opt.gpu self.forward_fn = forward_fn self.finished_epoch = 0 device = torch.device('cuda' if self.gpu else "cpu") self.model = self.model.to(device) if model_opt.resume is not None: self.set_model_weights(model_opt.resume) else: self.set_model_weights('kaiming') print_save(model_opt.logdir, self.model)
Example #17
Source File: model.py From OpenQA with MIT License | 5 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ logger.info("init_optimizer") if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False for p in self.selector.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] parameters = parameters + [p for p in self.selector.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #18
Source File: recommender.py From context_attentive_ir with MIT License | 5 votes |
def init_optimizer(self, state_dict=None, use_gpu=True): """Initialize an optimizer for the free parameters of the network. Args: state_dict: optimizer's state dict use_gpu: required to move state_dict to GPU """ if self.args.fix_embeddings: for p in self.network.embedder.word_embeddings.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adam': self.optimizer = optim.Adam(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adadelta': self.optimizer = optim.Adadelta(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) if state_dict is not None: self.optimizer.load_state_dict(state_dict) # FIXME: temp soln - https://github.com/pytorch/pytorch/issues/2830 if use_gpu: for state in self.optimizer.state.values(): for k, v in state.items(): if isinstance(v, torch.Tensor): state[k] = v.cuda() # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #19
Source File: multitask.py From context_attentive_ir with MIT License | 5 votes |
def init_optimizer(self, state_dict=None, use_gpu=True): """Initialize an optimizer for the free parameters of the network. Args: state_dict: optimizer's state dict use_gpu: required to move state_dict to GPU """ if self.args.fix_embeddings: for p in self.network.embedder.word_embeddings.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adam': self.optimizer = optim.Adam(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adadelta': self.optimizer = optim.Adadelta(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) if state_dict is not None: self.optimizer.load_state_dict(state_dict) # FIXME: temp soln - https://github.com/pytorch/pytorch/issues/2830 if use_gpu: for state in self.optimizer.state.values(): for k, v in state.items(): if isinstance(v, torch.Tensor): state[k] = v.cuda() # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #20
Source File: ranker.py From context_attentive_ir with MIT License | 5 votes |
def init_optimizer(self, state_dict=None, use_gpu=True): """Initialize an optimizer for the free parameters of the network. Args: state_dict: optimizer's state dict use_gpu: required to move state_dict to GPU """ if self.args.fix_embeddings: for p in self.network.word_embeddings.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adam': self.optimizer = optim.Adam(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adadelta': self.optimizer = optim.Adadelta(parameters, self.args.learning_rate, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) if state_dict is not None: self.optimizer.load_state_dict(state_dict) # FIXME: temp soln - https://github.com/pytorch/pytorch/issues/2830 if use_gpu: for state in self.optimizer.state.values(): for k, v in state.items(): if isinstance(v, torch.Tensor): state[k] = v.cuda() # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #21
Source File: model_CoQA.py From FlowDelta with MIT License | 5 votes |
def __init__(self, opt, embedding=None, state_dict=None): # Book-keeping. self.opt = opt self.updates = state_dict['updates'] if state_dict else 0 self.eval_embed_transfer = True self.train_loss = AverageMeter() # Building network. self.network = FlowQA(opt, embedding) if state_dict: new_state = set(self.network.state_dict().keys()) for k in list(state_dict['network'].keys()): if k not in new_state: del state_dict['network'][k] self.network.load_state_dict(state_dict['network']) # Building optimizer. parameters = [p for p in self.network.parameters() if p.requires_grad] if opt['optimizer'] == 'sgd': self.optimizer = optim.SGD(parameters, opt['learning_rate'], momentum=opt['momentum'], weight_decay=opt['weight_decay']) elif opt['optimizer'] == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=opt['weight_decay']) elif opt['optimizer'] == 'adadelta': self.optimizer = optim.Adadelta(parameters, rho=0.95, weight_decay=opt['weight_decay']) else: raise RuntimeError('Unsupported optimizer: %s' % opt['optimizer']) if state_dict: self.optimizer.load_state_dict(state_dict['optimizer']) if opt['fix_embeddings']: wvec_size = 0 else: wvec_size = (opt['vocab_size'] - opt['tune_partial']) * opt['embedding_dim'] self.total_param = sum([p.nelement() for p in parameters]) - wvec_size
Example #22
Source File: model.py From sru with MIT License | 5 votes |
def __init__(self, opt, embedding=None, state_dict=None): # Book-keeping. self.opt = opt self.updates = state_dict['updates'] if state_dict else 0 self.train_loss = AverageMeter() # Building network. self.network = RnnDocReader(opt, embedding=embedding) if state_dict: new_state = set(self.network.state_dict().keys()) for k in list(state_dict['network'].keys()): if k not in new_state: del state_dict['network'][k] self.network.load_state_dict(state_dict['network']) # Building optimizer. parameters = [p for p in self.network.parameters() if p.requires_grad] if opt['optimizer'] == 'sgd': self.optimizer = optim.SGD(parameters, opt['learning_rate'], momentum=opt['momentum'], weight_decay=opt['weight_decay']) elif opt['optimizer'] == 'adamax': self.optimizer = optim.Adamax(parameters, opt['learning_rate'], weight_decay=opt['weight_decay']) else: raise RuntimeError('Unsupported optimizer: %s' % opt['optimizer']) if state_dict: self.optimizer.load_state_dict(state_dict['optimizer']) num_params = sum(p.data.numel() for p in parameters if p.data.data_ptr() != self.network.embedding.weight.data.data_ptr()) print ("{} parameters".format(num_params))
Example #23
Source File: train_para_encoder.py From Multi-Step-Reasoning with Apache License 2.0 | 5 votes |
def init_from_checkpoint(args): logger.info('Loading model from saved checkpoint {}'.format(args.pretrained)) model = torch.load(args.pretrained) word_dict = model['word_dict'] feature_dict = model['feature_dict'] args.vocab_size = len(word_dict) args.embedding_dim_orig = args.embedding_dim args.word_dict = word_dict args.feature_dict = feature_dict ret = LSTMRetriever(args, word_dict, feature_dict) # load saved param values ret.model.load_state_dict(model['state_dict']['para_clf']) optimizer = None parameters = ret.get_trainable_params() if args.optimizer == 'sgd': optimizer = optim.SGD(parameters, args.learning_rate, momentum=args.momentum, weight_decay=args.weight_decay) elif args.optimizer == 'adamax': optimizer = optim.Adamax(parameters, weight_decay=args.weight_decay) elif args.optimizer == 'nag': optimizer = NAG(parameters, args.learning_rate, momentum=args.momentum, weight_decay=args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % args.optimizer) optimizer.load_state_dict(model['state_dict']['optimizer']) logger.info('Model loaded...') return ret, optimizer, word_dict, feature_dict
Example #24
Source File: train_para_encoder.py From Multi-Step-Reasoning with Apache License 2.0 | 5 votes |
def init_from_scratch(args, train_exs): logger.info('Initializing model from scratch') word_dict = feature_dict = None # create or get vocab word_dict = utils.build_word_dict(args, train_exs) if word_dict is not None: args.vocab_size = len(word_dict) args.embedding_dim_orig = args.embedding_dim args.word_dict = word_dict args.feature_dict = feature_dict ret = LSTMRetriever(args, word_dict, feature_dict) # -------------------------------------------------------------------------- # TRAIN/VALID LOOP # -------------------------------------------------------------------------- # train parameters = ret.get_trainable_params() optimizer = None if parameters is not None and len(parameters) > 0: if args.optimizer == 'sgd': optimizer = optim.SGD(parameters, args.learning_rate, momentum=args.momentum, weight_decay=args.weight_decay) elif args.optimizer == 'adamax': optimizer = optim.Adamax(parameters, weight_decay=args.weight_decay) elif args.optimizer == 'nag': optimizer = NAG(parameters, args.learning_rate, momentum=args.momentum, weight_decay=args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % args.optimizer) else: pass return ret, optimizer, word_dict, feature_dict
Example #25
Source File: model.py From Multi-Step-Reasoning with Apache License 2.0 | 5 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.multi_step_reasoner is not None: parameters += [p for p in self.multi_step_reasoner.parameters() if p.requires_grad] parameters += [p for p in self.reader_self_attn.parameters() if p.requires_grad] if self.multi_step_reader is not None: parameters += [p for p in self.multi_step_reader.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #26
Source File: model.py From commonsense-rc with MIT License | 5 votes |
def init_optimizer(self): parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, self.lr, momentum=0.4, weight_decay=0) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, lr=self.lr, weight_decay=0) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) self.scheduler = lr_scheduler.MultiStepLR(self.optimizer, milestones=[10, 15], gamma=0.5)
Example #27
Source File: utils.py From DeMa-BWE with BSD 3-Clause "New" or "Revised" License | 5 votes |
def get_optimizer(s): """ Parse optimizer parameters. Input should be of the form: - "sgd,lr=0.01" - "adagrad,lr=0.1,lr_decay=0.05" """ if "," in s: method = s[:s.find(',')] optim_params = {} for x in s[s.find(',') + 1:].split(','): split = x.split('=') assert len(split) == 2 assert re.match("^[+-]?(\d+(\.\d*)?|\.\d+)$", split[1]) is not None optim_params[split[0]] = float(split[1]) else: method = s optim_params = {} if method == 'adadelta': optim_fn = optim.Adadelta elif method == 'adagrad': optim_fn = optim.Adagrad elif method == 'adam': optim_fn = optim.Adam elif method == 'adamax': optim_fn = optim.Adamax elif method == 'asgd': optim_fn = optim.ASGD elif method == 'rmsprop': optim_fn = optim.RMSprop elif method == 'rprop': optim_fn = optim.Rprop elif method == 'sgd': optim_fn = optim.SGD assert 'lr' in optim_params else: raise Exception('Unknown optimization method: "%s"' % method) return optim_fn, optim_params
Example #28
Source File: model.py From neural_chat with MIT License | 5 votes |
def __init__(self, opt, word_dict, feature_dict, state_dict=None): # Book-keeping. self.opt = opt self.word_dict = word_dict self.feature_dict = feature_dict self.updates = 0 self.train_loss = AverageMeter() # Building network. self.network = RnnDocReader(opt) if state_dict: new_state = set(self.network.state_dict().keys()) for k in list(state_dict['network'].keys()): if k not in new_state: del state_dict['network'][k] self.network.load_state_dict(state_dict['network']) # Building optimizer. parameters = [p for p in self.network.parameters() if p.requires_grad] if opt['optimizer'] == 'sgd': self.optimizer = optim.SGD( parameters, opt['learning_rate'], momentum=opt['momentum'], weight_decay=opt['weight_decay'], ) elif opt['optimizer'] == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=opt['weight_decay']) else: raise RuntimeError('Unsupported optimizer: %s' % opt['optimizer'])
Example #29
Source File: model.py From MnemonicReader with BSD 3-Clause "New" or "Revised" License | 5 votes |
def init_optimizer(self, state_dict=None): """Initialize an optimizer for the free parameters of the network. Args: state_dict: network parameters """ if self.args.fix_embeddings: for p in self.network.embedding.parameters(): p.requires_grad = False parameters = [p for p in self.network.parameters() if p.requires_grad] if self.args.optimizer == 'sgd': self.optimizer = optim.SGD(parameters, lr=self.args.learning_rate, momentum=self.args.momentum, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adamax': self.optimizer = optim.Adamax(parameters, weight_decay=self.args.weight_decay) elif self.args.optimizer == 'adadelta': self.optimizer = optim.Adadelta(parameters, lr=self.args.learning_rate, rho=self.args.rho, eps=self.args.eps, weight_decay=self.args.weight_decay) else: raise RuntimeError('Unsupported optimizer: %s' % self.args.optimizer) # -------------------------------------------------------------------------- # Learning # --------------------------------------------------------------------------
Example #30
Source File: get_optimizer.py From PyMIC with Apache License 2.0 | 5 votes |
def get_optimiser(name, net_params, optim_params): lr = optim_params['learning_rate'] momentum = optim_params['momentum'] weight_decay = optim_params['weight_decay'] if(name == "SGD"): return optim.SGD(net_params, lr, momentum = momentum, weight_decay = weight_decay) elif(name == "Adam"): return optim.Adam(net_params, lr, weight_decay = 1e-5) elif(name == "SparseAdam"): return optim.SparseAdam(net_params, lr) elif(name == "Adadelta"): return optim.Adadelta(net_params, lr, weight_decay = weight_decay) elif(name == "Adagrad"): return optim.Adagrad(net_params, lr, weight_decay = weight_decay) elif(name == "Adamax"): return optim.Adamax(net_params, lr, weight_decay = weight_decay) elif(name == "ASGD"): return optim.ASGD(net_params, lr, weight_decay = weight_decay) elif(name == "LBFGS"): return optim.LBFGS(net_params, lr) elif(name == "RMSprop"): return optim.RMSprop(net_params, lr, momentum = momentum, weight_decay = weight_decay) elif(name == "Rprop"): return optim.Rprop(net_params, lr) else: raise ValueError("unsupported optimizer {0:}".format(name))