12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061 |
- from __future__ import absolute_import
- from __future__ import division
- from __future__ import print_function
- from __future__ import unicode_literals
- import copy
- import paddle
- __all__ = ['build_optimizer']
- def build_lr_scheduler(lr_config, epochs, step_each_epoch):
- from . import learning_rate
- lr_config.update({'epochs': epochs, 'step_each_epoch': step_each_epoch})
- if 'name' in lr_config:
- lr_name = lr_config.pop('name')
- lr = getattr(learning_rate, lr_name)(**lr_config)()
- else:
- lr = lr_config['learning_rate']
- return lr
- def build_optimizer(config, epochs, step_each_epoch, parameters):
- from . import regularizer, optimizer
- config = copy.deepcopy(config)
-
- lr = build_lr_scheduler(config.pop('lr'), epochs, step_each_epoch)
-
- if 'regularizer' in config and config['regularizer'] is not None:
- reg_config = config.pop('regularizer')
- reg_name = reg_config.pop('name') + 'Decay'
- reg = getattr(regularizer, reg_name)(**reg_config)()
- else:
- reg = None
-
- optim_name = config.pop('name')
- if 'clip_norm' in config:
- clip_norm = config.pop('clip_norm')
- grad_clip = paddle.nn.ClipGradByNorm(clip_norm=clip_norm)
- else:
- grad_clip = None
- optim = getattr(optimizer, optim_name)(learning_rate=lr,
- weight_decay=reg,
- grad_clip=grad_clip,
- **config)
- return optim(parameters), lr
|