13 #ifndef MLPACK_METHODS_RL_TRAINING_CONFIG_HPP 14 #define MLPACK_METHODS_RL_TRAINING_CONFIG_HPP 28 doubleQLearning(false)
33 size_t updateInterval,
34 size_t targetNetworkSyncInterval,
36 size_t explorationSteps,
40 bool doubleQLearning) :
41 numWorkers(numWorkers),
42 updateInterval(updateInterval),
43 targetNetworkSyncInterval(targetNetworkSyncInterval),
45 explorationSteps(explorationSteps),
48 gradientLimit(gradientLimit),
49 doubleQLearning(doubleQLearning)
64 {
return targetNetworkSyncInterval; }
114 size_t updateInterval;
120 size_t targetNetworkSyncInterval;
133 size_t explorationSteps;
151 double gradientLimit;
157 bool doubleQLearning;
double & Discount()
Modify the discount rate for future reward.
size_t ExplorationSteps() const
Get the exploration steps.
size_t NumWorkers() const
Get the amount of workers.
TrainingConfig(size_t numWorkers, size_t updateInterval, size_t targetNetworkSyncInterval, size_t stepLimit, size_t explorationSteps, double stepSize, double discount, double gradientLimit, bool doubleQLearning)
size_t StepLimit() const
Get the maximum steps of each episode.
size_t TargetNetworkSyncInterval() const
Get the interval for syncing target network.
size_t & StepLimit()
Modify the maximum steps of each episode.
size_t & TargetNetworkSyncInterval()
Modify the interval for syncing target network.
size_t UpdateInterval() const
Get the update interval.
double Discount() const
Get the discount rate for future reward.
size_t & NumWorkers()
Modify the amount of workers.
bool DoubleQLearning() const
Get the indicator of double q-learning.
size_t & ExplorationSteps()
Modify the exploration steps.
bool & DoubleQLearning()
Modify the indicator of double q-learning.
double GradientLimit() const
Get the limit of update gradient.
size_t & UpdateInterval()
Modify the update interval.
double & GradientLimit()
Modify the limit of update gradient.
double StepSize() const
Get the step size of the optimizer.
double & StepSize()
Modify the step size of the optimizer.