Skip to content

Commit

Permalink
remove deprecated tp comm overlap configs
Browse files Browse the repository at this point in the history
Signed-off-by: Sangkug Lym <[email protected]>
  • Loading branch information
erhoo82 committed Jul 31, 2024
1 parent 95c77b5 commit 1a7171d
Show file tree
Hide file tree
Showing 34 changed files with 0 additions and 70 deletions.
2 changes: 0 additions & 2 deletions launcher_scripts/conf/peft/llama/sft.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,8 +76,6 @@ model:
sync_batch_comm: False
overlap_p2p_comm: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

megatron_amp_O2: True
mcore_gpt: True
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/peft/nemotron/sft.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,8 +76,6 @@ model:
sync_batch_comm: False
overlap_p2p_comm: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

megatron_amp_O2: True
mcore_gpt: True
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/peft/qwen2/sft.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,8 +76,6 @@ model:
sync_batch_comm: False
overlap_p2p_comm: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

megatron_amp_O2: True
mcore_gpt: True
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/peft/starcoder2/sft.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,8 +76,6 @@ model:
sync_batch_comm: False
overlap_p2p_comm: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

megatron_amp_O2: True
mcore_gpt: True
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/126m.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -144,8 +144,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/175b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/175b_16k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/175b_32k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/175b_fp8.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/1b_improved.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -150,8 +150,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

optim:
name: distributed_fused_adam
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/20b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/400m_improved.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -150,8 +150,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

optim:
name: distributed_fused_adam
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/40b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/40b_16k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/40b_32k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/40b_64k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/40b_improved.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -150,8 +150,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

optim:
name: distributed_fused_adam
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/5b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/5b_16k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/5b_32k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/5b_64k.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

# miscellaneous
seed: 1234
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/gpt3/7b_improved.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -150,8 +150,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

optim:
name: distributed_fused_adam
Expand Down
4 changes: 0 additions & 4 deletions launcher_scripts/conf/training/gpt3/mlperf.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -268,7 +268,3 @@ model:
gc_interval: 100
name: megatron_gpt_full_te_layer_autocast
fp8_params: true
tp_comm_split_ag: true
tp_comm_split_rs: false
tp_comm_atomic_ag: false
tp_comm_atomic_rs: true
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/llama/llama2_13b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -136,8 +136,6 @@ model:
ub_tp_comm_overlap: false
overlap_p2p_comm: true
batch_p2p_comm: false
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
nsys_profile:
enabled: False
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/llama/llama2_70b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -134,8 +134,6 @@ model:
fp8_amax_compute_algo: most_recent
use_emha: false
ub_tp_comm_overlap: true
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
overlap_p2p_comm: true
batch_p2p_comm: false
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/llama/llama2_7b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -136,8 +136,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
use_emha: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
nsys_profile:
enabled: False
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/nemotron/nemotron_15b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -156,8 +156,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

nsys_profile:
enabled: False
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/nemotron/nemotron_22b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -156,8 +156,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: True
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

gc_interval: 100

Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/nemotron/nemotron_4b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -156,8 +156,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: true
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

nsys_profile:
enabled: False
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/nemotron/nemotron_8b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -156,8 +156,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
fp8_wgrad: True
ub_tp_comm_overlap: true
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False

nsys_profile:
enabled: False
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/qwen2/qwen2_14b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -138,8 +138,6 @@ model:
fp8_amax_compute_algo: most_recent
use_emha: false
ub_tp_comm_overlap: true
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
optim:
name: distributed_fused_adam
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/qwen2/qwen2_4b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -140,8 +140,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
use_emha: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
optim:
name: distributed_fused_adam
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/qwen2/qwen2_72b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -137,8 +137,6 @@ model:
fp8_amax_compute_algo: most_recent
use_emha: false
ub_tp_comm_overlap: true
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
overlap_p2p_comm: true
batch_p2p_comm: false
Expand Down
2 changes: 0 additions & 2 deletions launcher_scripts/conf/training/qwen2/qwen2_7b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -140,8 +140,6 @@ model:
fp8_amax_compute_algo: max # 'most_recent' or 'max'. Algorithm for computing amax from history
use_emha: False
ub_tp_comm_overlap: False
tp_comm_atomic_ag: False
tp_comm_atomic_rs: False
use_flash_attention: true
optim:
name: distributed_fused_adam
Expand Down

0 comments on commit 1a7171d

Please sign in to comment.