Source code for d3rlpy.algos.cql

from typing import Any, List, Optional, Sequence
from .base import AlgoBase, DataGenerator
from .dqn import DoubleDQN
from .torch.cql_impl import CQLImpl, DiscreteCQLImpl
from ..augmentation import AugmentationPipeline
from ..dataset import TransitionMiniBatch
from ..models.encoders import EncoderFactory
from ..models.q_functions import QFunctionFactory
from ..models.optimizers import OptimizerFactory, AdamFactory
from ..gpu import Device
from ..argument_utility import check_encoder, EncoderArg
from ..argument_utility import check_use_gpu, UseGPUArg
from ..argument_utility import check_augmentation, AugmentationArg
from ..argument_utility import check_q_func, QFuncArg
from ..argument_utility import ScalerArg, ActionScalerArg
from ..constants import IMPL_NOT_INITIALIZED_ERROR


[docs]class CQL(AlgoBase): r"""Conservative Q-Learning algorithm. CQL is a SAC-based data-driven deep reinforcement learning algorithm, which achieves state-of-the-art performance in offline RL problems. CQL mitigates overestimation error by minimizing action-values under the current policy and maximizing values under data distribution for underestimation issue. .. math:: L(\theta_i) = \alpha\, \mathbb{E}_{s_t \sim D} \left[\log{\sum_a \exp{Q_{\theta_i}(s_t, a)}} - \mathbb{E}_{a \sim D} \big[Q_{\theta_i}(s, a)\big] - \tau\right] + L_\mathrm{SAC}(\theta_i) where :math:`\alpha` is an automatically adjustable value via Lagrangian dual gradient descent and :math:`\tau` is a threshold value. If the action-value difference is smaller than :math:`\tau`, the :math:`\alpha` will become smaller. Otherwise, the :math:`\alpha` will become larger to aggressively penalize action-values. In continuous control, :math:`\log{\sum_a \exp{Q(s, a)}}` is computed as follows. .. math:: \log{\sum_a \exp{Q(s, a)}} \approx \log{\left( \frac{1}{2N} \sum_{a_i \sim \text{Unif}(a)}^N \left[\frac{\exp{Q(s, a_i)}}{\text{Unif}(a)}\right] + \frac{1}{2N} \sum_{a_i \sim \pi_\phi(a|s)}^N \left[\frac{\exp{Q(s, a_i)}}{\pi_\phi(a_i|s)}\right]\right)} where :math:`N` is the number of sampled actions. The rest of optimization is exactly same as :class:`d3rlpy.algos.SAC`. References: * `Kumar et al., Conservative Q-Learning for Offline Reinforcement Learning. <https://arxiv.org/abs/2006.04779>`_ Args: actor_learning_rate (float): learning rate for policy function. critic_learning_rate (float): learning rate for Q functions. temp_learning_rate (float): learning rate for temperature parameter of SAC. alpha_learning_rate (float): learning rate for :math:`\alpha`. actor_optim_factory (d3rlpy.models.optimizers.OptimizerFactory): optimizer factory for the actor. critic_optim_factory (d3rlpy.models.optimizers.OptimizerFactory): optimizer factory for the critic. temp_optim_factory (d3rlpy.models.optimizers.OptimizerFactory): optimizer factory for the temperature. alpha_optim_factory (d3rlpy.models.optimizers.OptimizerFactory): optimizer factory for :math:`\alpha`. actor_encoder_factory (d3rlpy.models.encoders.EncoderFactory or str): encoder factory for the actor. critic_encoder_factory (d3rlpy.models.encoders.EncoderFactory or str): encoder factory for the critic. q_func_factory (d3rlpy.models.q_functions.QFunctionFactory or str): Q function factory. batch_size (int): mini-batch size. n_frames (int): the number of frames to stack for image observation. n_steps (int): N-step TD calculation. gamma (float): discount factor. tau (float): target network synchronization coefficiency. n_critics (int): the number of Q functions for ensemble. bootstrap (bool): flag to bootstrap Q functions. share_encoder (bool): flag to share encoder network. target_reduction_type (str): ensemble reduction method at target value estimation. The available options are ``['min', 'max', 'mean', 'mix', 'none']``. update_actor_interval (int): interval to update policy function. initial_temperature (float): initial temperature value. initial_alpha (float): initial :math:`\alpha` value. alpha_threshold (float): threshold value described as :math:`\tau`. n_action_samples (int): the number of sampled actions to compute :math:`\log{\sum_a \exp{Q(s, a)}}`. soft_q_backup (bool): flag to use SAC-style backup. use_gpu (bool, int or d3rlpy.gpu.Device): flag to use GPU, device ID or device. scaler (d3rlpy.preprocessing.Scaler or str): preprocessor. The available options are `['pixel', 'min_max', 'standard']`. action_scaler (d3rlpy.preprocessing.ActionScaler or str): action preprocessor. The available options are ``['min_max']``. augmentation (d3rlpy.augmentation.AugmentationPipeline or list(str)): augmentation pipeline. generator (d3rlpy.algos.base.DataGenerator): dynamic dataset generator (e.g. model-based RL). impl (d3rlpy.algos.torch.cql_impl.CQLImpl): algorithm implementation. """ _actor_learning_rate: float _critic_learning_rate: float _temp_learning_rate: float _alpha_learning_rate: float _actor_optim_factory: OptimizerFactory _critic_optim_factory: OptimizerFactory _temp_optim_factory: OptimizerFactory _alpha_optim_factory: OptimizerFactory _actor_encoder_factory: EncoderFactory _critic_encoder_factory: EncoderFactory _q_func_factory: QFunctionFactory _tau: float _bootstrap: bool _n_critics: int _share_encoder: bool _target_reduction_type: str _update_actor_interval: int _initial_temperature: float _initial_alpha: float _alpha_threshold: float _n_action_samples: int _soft_q_backup: bool _augmentation: AugmentationPipeline _use_gpu: Optional[Device] _impl: Optional[CQLImpl] def __init__( self, *, actor_learning_rate: float = 1e-4, critic_learning_rate: float = 3e-4, temp_learning_rate: float = 1e-4, alpha_learning_rate: float = 1e-4, actor_optim_factory: OptimizerFactory = AdamFactory(), critic_optim_factory: OptimizerFactory = AdamFactory(), temp_optim_factory: OptimizerFactory = AdamFactory(), alpha_optim_factory: OptimizerFactory = AdamFactory(), actor_encoder_factory: EncoderArg = "default", critic_encoder_factory: EncoderArg = "default", q_func_factory: QFuncArg = "mean", batch_size: int = 256, n_frames: int = 1, n_steps: int = 1, gamma: float = 0.99, tau: float = 0.005, n_critics: int = 2, bootstrap: bool = False, share_encoder: bool = False, target_reduction_type: str = "min", update_actor_interval: int = 1, initial_temperature: float = 1.0, initial_alpha: float = 5.0, alpha_threshold: float = 10.0, n_action_samples: int = 10, soft_q_backup: bool = False, use_gpu: UseGPUArg = False, scaler: ScalerArg = None, action_scaler: ActionScalerArg = None, augmentation: AugmentationArg = None, generator: Optional[DataGenerator] = None, impl: Optional[CQLImpl] = None, **kwargs: Any ): super().__init__( batch_size=batch_size, n_frames=n_frames, n_steps=n_steps, gamma=gamma, scaler=scaler, action_scaler=action_scaler, generator=generator, ) self._actor_learning_rate = actor_learning_rate self._critic_learning_rate = critic_learning_rate self._temp_learning_rate = temp_learning_rate self._alpha_learning_rate = alpha_learning_rate self._actor_optim_factory = actor_optim_factory self._critic_optim_factory = critic_optim_factory self._temp_optim_factory = temp_optim_factory self._alpha_optim_factory = alpha_optim_factory self._actor_encoder_factory = check_encoder(actor_encoder_factory) self._critic_encoder_factory = check_encoder(critic_encoder_factory) self._q_func_factory = check_q_func(q_func_factory) self._tau = tau self._bootstrap = bootstrap self._n_critics = n_critics self._share_encoder = share_encoder self._target_reduction_type = target_reduction_type self._update_actor_interval = update_actor_interval self._initial_temperature = initial_temperature self._initial_alpha = initial_alpha self._alpha_threshold = alpha_threshold self._n_action_samples = n_action_samples self._soft_q_backup = soft_q_backup self._augmentation = check_augmentation(augmentation) self._use_gpu = check_use_gpu(use_gpu) self._impl = impl
[docs] def create_impl( self, observation_shape: Sequence[int], action_size: int ) -> None: self._impl = CQLImpl( observation_shape=observation_shape, action_size=action_size, actor_learning_rate=self._actor_learning_rate, critic_learning_rate=self._critic_learning_rate, temp_learning_rate=self._temp_learning_rate, alpha_learning_rate=self._alpha_learning_rate, actor_optim_factory=self._actor_optim_factory, critic_optim_factory=self._critic_optim_factory, temp_optim_factory=self._temp_optim_factory, alpha_optim_factory=self._alpha_optim_factory, actor_encoder_factory=self._actor_encoder_factory, critic_encoder_factory=self._critic_encoder_factory, q_func_factory=self._q_func_factory, gamma=self._gamma, tau=self._tau, n_critics=self._n_critics, bootstrap=self._bootstrap, share_encoder=self._share_encoder, target_reduction_type=self._target_reduction_type, initial_temperature=self._initial_temperature, initial_alpha=self._initial_alpha, alpha_threshold=self._alpha_threshold, n_action_samples=self._n_action_samples, soft_q_backup=self._soft_q_backup, use_gpu=self._use_gpu, scaler=self._scaler, action_scaler=self._action_scaler, augmentation=self._augmentation, ) self._impl.build()
[docs] def update( self, epoch: int, total_step: int, batch: TransitionMiniBatch ) -> List[Optional[float]]: assert self._impl is not None, IMPL_NOT_INITIALIZED_ERROR critic_loss = self._impl.update_critic( batch.observations, batch.actions, batch.next_rewards, batch.next_observations, batch.terminals, batch.n_steps, batch.masks, ) if total_step % self._update_actor_interval == 0: actor_loss = self._impl.update_actor(batch.observations) # lagrangian parameter update for SAC temperature if self._temp_learning_rate > 0: temp_loss, temp = self._impl.update_temp(batch.observations) else: temp_loss, temp = None, None # lagrangian parameter update for conservative loss weight if self._alpha_learning_rate > 0: alpha_loss, alpha = self._impl.update_alpha( batch.observations, batch.actions ) else: alpha_loss, alpha = None, None self._impl.update_critic_target() self._impl.update_actor_target() else: actor_loss = None temp_loss = None temp = None alpha_loss = None alpha = None return [critic_loss, actor_loss, temp_loss, temp, alpha_loss, alpha]
[docs] def get_loss_labels(self) -> List[str]: return [ "critic_loss", "actor_loss", "temp_loss", "temp", "alpha_loss", "alpha", ]
[docs]class DiscreteCQL(DoubleDQN): r"""Discrete version of Conservative Q-Learning algorithm. Discrete version of CQL is a DoubleDQN-based data-driven deep reinforcement learning algorithm (the original paper uses DQN), which achieves state-of-the-art performance in offline RL problems. CQL mitigates overestimation error by minimizing action-values under the current policy and maximizing values under data distribution for underestimation issue. .. math:: L(\theta) = \mathbb{E}_{s_t \sim D} [\log{\sum_a \exp{Q_{\theta}(s_t, a)}} - \mathbb{E}_{a \sim D} [Q_{\theta}(s, a)]] + L_{DoubleDQN}(\theta) References: * `Kumar et al., Conservative Q-Learning for Offline Reinforcement Learning. <https://arxiv.org/abs/2006.04779>`_ Args: learning_rate (float): learning rate. optim_factory (d3rlpy.models.optimizers.OptimizerFactory): optimizer factory. encoder_factory (d3rlpy.models.encoders.EncoderFactory or str): encoder factory. q_func_factory (d3rlpy.models.q_functions.QFunctionFactory or str): Q function factory. batch_size (int): mini-batch size. n_frames (int): the number of frames to stack for image observation. n_steps (int): N-step TD calculation. gamma (float): discount factor. n_critics (int): the number of Q functions for ensemble. bootstrap (bool): flag to bootstrap Q functions. share_encoder (bool): flag to share encoder network. target_reduction_type (str): ensemble reduction method at target value estimation. The available options are ``['min', 'max', 'mean', 'mix', 'none']``. target_update_interval (int): interval to synchronize the target network. use_gpu (bool, int or d3rlpy.gpu.Device): flag to use GPU, device ID or device. scaler (d3rlpy.preprocessing.Scaler or str): preprocessor. The available options are `['pixel', 'min_max', 'standard']` augmentation (d3rlpy.augmentation.AugmentationPipeline or list(str)): augmentation pipeline. generator (d3rlpy.algos.base.DataGenerator): dynamic dataset generator (e.g. model-based RL). impl (d3rlpy.algos.torch.cql_impl.DiscreteCQLImpl): algorithm implementation. """ _impl: Optional[DiscreteCQLImpl]
[docs] def create_impl( self, observation_shape: Sequence[int], action_size: int ) -> None: self._impl = DiscreteCQLImpl( observation_shape=observation_shape, action_size=action_size, learning_rate=self._learning_rate, optim_factory=self._optim_factory, encoder_factory=self._encoder_factory, q_func_factory=self._q_func_factory, gamma=self._gamma, n_critics=self._n_critics, bootstrap=self._bootstrap, share_encoder=self._share_encoder, target_reduction_type=self._target_reduction_type, use_gpu=self._use_gpu, scaler=self._scaler, augmentation=self._augmentation, ) self._impl.build()