diff --git a/examples/automated_deep_compression/presets/ADC_ClippedPPO.py b/examples/automated_deep_compression/presets/ADC_ClippedPPO.py index 464155bac90572a07cf4a88a207a8acb1cdbd11f..4f1e788122d1d77c4391b4e1aeb22daa96da0632 100755 --- a/examples/automated_deep_compression/presets/ADC_ClippedPPO.py +++ b/examples/automated_deep_compression/presets/ADC_ClippedPPO.py @@ -1,10 +1,9 @@ from rl_coach.agents.clipped_ppo_agent import ClippedPPOAgentParameters from rl_coach.architectures.layers import Dense -from rl_coach.base_parameters import VisualizationParameters, PresetValidationParameters, DistributedCoachSynchronizationType -from rl_coach.core_types import TrainingSteps, EnvironmentEpisodes, EnvironmentSteps, RunPhase -from rl_coach.environments.gym_environment import GymVectorEnvironment, mujoco_v2 -from rl_coach.exploration_policies.additive_noise import AdditiveNoiseParameters -from rl_coach.exploration_policies.e_greedy import EGreedyParameters +from rl_coach.base_parameters import VisualizationParameters, DistributedCoachSynchronizationType +from rl_coach.core_types import TrainingSteps, EnvironmentEpisodes, EnvironmentSteps +from rl_coach.environments.gym_environment import GymVectorEnvironment +from rl_coach.architectures.head_parameters import PPOHeadParameters, VHeadParameters from rl_coach.filters.filter import InputFilter from rl_coach.filters.observation.observation_normalization_filter import ObservationNormalizationFilter from rl_coach.graph_managers.basic_rl_graph_manager import BasicRLGraphManager