@@ -824,21 +824,21 @@ Setting any of the following options selects `Dusty's annealing schedule <dusty_
824
824
.. option :: --place_cost_exp <float >
825
825
826
826
Wiring cost is divided by the average channel width over a net's bounding box
827
- taken to this exponent.Only impacts devices with different channel widths in
827
+ taken to this exponent. Only impacts devices with different channel widths in
828
828
different directions or regions.
829
829
830
830
**Default: ** ``1 ``
831
831
832
832
.. option :: --RL_agent_placement {on | off}
833
833
834
834
Uses a Reinforcement Learning (RL) agent in choosing the appropiate move type in placement.
835
- It activates the RL agent placement instead of using fixed probability for each move type.
835
+ It activates the RL agent placement instead of using a fixed probability for each move type.
836
836
837
837
**Default: ** ``on ``
838
838
839
839
.. option :: --place_agent_multistate {on | off}
840
840
841
- Enable multistate agent in the placement. A second state will be activated late in
841
+ Enable a multistate agent in the placement. A second state will be activated late in
842
842
the annealing and in the Quench that includes all the timing driven directed moves.
843
843
844
844
**Default: ** ``on ``
@@ -851,7 +851,7 @@ Setting any of the following options selects `Dusty's annealing schedule <dusty_
851
851
852
852
.. option :: --place_agent_epsilon <float >
853
853
854
- Placement RL agent's epsilon for epsilon-greedy agent. Epsilon represents
854
+ Placement RL agent's epsilon for the epsilon-greedy agent. Epsilon represents
855
855
the percentage of exploration actions taken vs the exploitation ones.
856
856
857
857
**Default: ** ``0.3 ``
@@ -867,16 +867,15 @@ Setting any of the following options selects `Dusty's annealing schedule <dusty_
867
867
868
868
.. option :: --place_reward_fun {basic | nonPenalizing_basic | runtime_aware | WLbiased_runtime_aware}
869
869
870
- The reward function used by placement RL agent to learn best action at each anneal stage.
870
+ The reward function used by the placement RL agent to learn the best action at each anneal stage.
871
871
872
872
.. note :: The latter two are only available for timing-driven placement.
873
873
874
874
**Default: ** ``WLbiased_runtime_aware ``
875
875
876
876
.. option :: --place_agent_space {move_type | move_block_type}
877
877
878
- Agent exploration space can be either based on only move types or also consider different block types.
879
- The available values are: move_type, move_block_type
878
+ The RL Agent exploration space can be either based on only move types or also consider different block types moved.
880
879
881
880
**Default: ** ``move_block_type ``
882
881
0 commit comments