diff --git a/examples/automated_deep_compression/presets/ADC_ClippedPPO.py b/examples/automated_deep_compression/presets/ADC_ClippedPPO.py
index 464155bac90572a07cf4a88a207a8acb1cdbd11f..4f1e788122d1d77c4391b4e1aeb22daa96da0632 100755
--- a/examples/automated_deep_compression/presets/ADC_ClippedPPO.py
+++ b/examples/automated_deep_compression/presets/ADC_ClippedPPO.py
@@ -1,10 +1,9 @@
 from rl_coach.agents.clipped_ppo_agent import ClippedPPOAgentParameters
 from rl_coach.architectures.layers import Dense
-from rl_coach.base_parameters import VisualizationParameters, PresetValidationParameters, DistributedCoachSynchronizationType
-from rl_coach.core_types import TrainingSteps, EnvironmentEpisodes, EnvironmentSteps, RunPhase
-from rl_coach.environments.gym_environment import GymVectorEnvironment, mujoco_v2
-from rl_coach.exploration_policies.additive_noise import AdditiveNoiseParameters
-from rl_coach.exploration_policies.e_greedy import EGreedyParameters
+from rl_coach.base_parameters import VisualizationParameters, DistributedCoachSynchronizationType
+from rl_coach.core_types import TrainingSteps, EnvironmentEpisodes, EnvironmentSteps
+from rl_coach.environments.gym_environment import GymVectorEnvironment
+from rl_coach.architectures.head_parameters import PPOHeadParameters, VHeadParameters
 from rl_coach.filters.filter import InputFilter
 from rl_coach.filters.observation.observation_normalization_filter import ObservationNormalizationFilter
 from rl_coach.graph_managers.basic_rl_graph_manager import BasicRLGraphManager