Skip to content

Commit

Permalink
Merge pull request #716 from mlcommons/variants_target_setting_clean
Browse files Browse the repository at this point in the history
Add configs for variants target setting.
  • Loading branch information
priyakasimbeg authored Mar 22, 2024
2 parents d0ed25a + dcc14d7 commit 231fc5b
Show file tree
Hide file tree
Showing 39 changed files with 1,234 additions and 32 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -272,11 +272,11 @@ def use_silu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22009
return 1 - 0.22009

@property
def test_target_value(self) -> float:
return 0.3426
return 1 - 0.3426


class ImagenetResNetGELUWorkload(ImagenetResNetWorkload):
Expand All @@ -287,11 +287,11 @@ def use_gelu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22077
return 1 - 0.22077

@property
def test_target_value(self) -> float:
return 0.3402
return 1 - 0.3402


class ImagenetResNetLargeBNScaleWorkload(ImagenetResNetWorkload):
Expand All @@ -302,8 +302,8 @@ def bn_init_scale(self) -> float:

@property
def validation_target_value(self) -> float:
return 0.23474
return 1 - 0.23474

@property
def test_target_value(self) -> float:
return 0.3577
return 1 - 0.3577
Original file line number Diff line number Diff line change
Expand Up @@ -326,11 +326,11 @@ def use_silu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22009
return 1 - 0.22009

@property
def test_target_value(self) -> float:
return 0.342
return 1 - 0.342


class ImagenetResNetGELUWorkload(ImagenetResNetWorkload):
Expand All @@ -341,11 +341,11 @@ def use_gelu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22077
return 1 - 0.22077

@property
def test_target_value(self) -> float:
return 0.3402
return 1 - 0.3402


class ImagenetResNetLargeBNScaleWorkload(ImagenetResNetWorkload):
Expand All @@ -356,8 +356,8 @@ def bn_init_scale(self) -> float:

@property
def validation_target_value(self) -> float:
return 0.23474
return 1 - 0.23474

@property
def test_target_value(self) -> float:
return 0.3577
return 1 - 0.3577
Original file line number Diff line number Diff line change
Expand Up @@ -99,11 +99,11 @@ def use_glu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.2233
return 1 - 0.2233

@property
def test_target_value(self) -> float:
return 0.3455
return 1 - 0.3455


class ImagenetVitPostLNWorkload(ImagenetVitWorkload):
Expand All @@ -114,11 +114,11 @@ def use_post_layer_norm(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.24688
return 1 - 0.24688

@property
def test_target_value(self) -> float:
return 0.3714
return 1 - 0.3714


class ImagenetVitMapWorkload(ImagenetVitWorkload):
Expand All @@ -129,8 +129,8 @@ def use_map(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22886
return 1 - 0.22886

@property
def test_target_value(self) -> float:
return 0.3477
return 1 - 0.3477
Original file line number Diff line number Diff line change
Expand Up @@ -90,11 +90,11 @@ def use_glu(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.2233
return 1 - 0.2233

@property
def test_target_value(self) -> float:
return 0.3455
return 1 - 0.3455


class ImagenetVitPostLNWorkload(ImagenetVitWorkload):
Expand All @@ -105,11 +105,11 @@ def use_post_layer_norm(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.24688
return 1 - 0.24688

@property
def test_target_value(self) -> float:
return 0.3714
return 1 - 0.3714


class ImagenetVitMapWorkload(ImagenetVitWorkload):
Expand All @@ -120,8 +120,8 @@ def use_map(self) -> bool:

@property
def validation_target_value(self) -> float:
return 0.22886
return 1 - 0.22886

@property
def test_target_value(self) -> float:
return 0.3477
return 1 - 0.3477
Original file line number Diff line number Diff line change
Expand Up @@ -107,13 +107,29 @@ class LibriSpeechDeepSpeechTanhWorkload(LibriSpeechDeepSpeechWorkload):
def use_tanh(self) -> bool:
return True

@property
def validation_target_value(self) -> float:
return 0.133449

@property
def test_target_value(self) -> float:
return 0.079810


class LibriSpeechDeepSpeechNoResNetWorkload(LibriSpeechDeepSpeechWorkload):

@property
def enable_residual_connections(self) -> bool:
return False

@property
def validation_target_value(self) -> float:
return 0.105042

@property
def test_target_value(self) -> float:
return 0.060388


class LibriSpeechDeepSpeechNormAndSpecAugWorkload(LibriSpeechDeepSpeechWorkload
):
Expand All @@ -137,3 +153,11 @@ def freq_mask_count(self) -> int:
@property
def time_mask_count(self) -> int:
return 15

@property
def validation_target_value(self) -> float:
return 0.131553

@property
def test_target_value(self) -> float:
return 0.082442
2 changes: 1 addition & 1 deletion algorithmic_efficiency/workloads/workloads.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@
},
'imagenet_vit_map': {
'workload_path': 'imagenet_vit/imagenet',
'workload_class_name': 'ImagenetVitMapLNWorkload',
'workload_class_name': 'ImagenetVitMapWorkload',
},
'librispeech_conformer': {
'workload_path': 'librispeech_conformer/librispeech',
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
{
"learning_rate": {
"feasible_points": [
0.002517072211464665
]
},
"beta1": {
"feasible_points": [
0.9908351643533544
]
},
"beta2": {
"feasible_points": [
0.9859568907533993
]
},
"warmup_steps": {
"feasible_points": [
799
]
},
"weight_decay": {
"feasible_points": [
0.12274552870237089
]
}
}

Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
{
"learning_rate": {
"feasible_points": [
0.05493199486120455
]
},
"beta1": {
"feasible_points": [
0.954922991734919
]
},
"beta2": {
"feasible_points": [
0.9986188074995163
]
},
"warmup_steps": {
"feasible_points": [
799
]
},
"weight_decay": {
"feasible_points": [
0.00011065469792077193
]
}
}

Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
{
"learning_rate": {
"feasible_points": [
0.001493629901423942
]
},
"beta1": {
"feasible_points": [
0.9592129978682067
]
},
"beta2": {
"feasible_points": [
0.9824918272399145
]
},
"warmup_steps": {
"feasible_points": [
399
]
},
"weight_decay": {
"feasible_points": [
0.00038587516415285595
]
}
}

Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
{
"learning_rate": {
"feasible_points": [
0.008334676559764446
]
},
"beta1": {
"feasible_points": [
0.8294338711079317
]
},
"beta2": {
"feasible_points": [
0.8551723332825868
]
},
"warmup_steps": {
"feasible_points": [
2714
]
},
"weight_decay": {
"feasible_points": [
0.01371235755699044
]
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
{
"learning_rate": {
"feasible_points": [
0.006173154695175443
]
},
"beta1": {
"feasible_points": [
0.8496694604806512
]
},
"beta2": {
"feasible_points": [
0.4639437428687345
]
},
"warmup_steps": {
"feasible_points": [
1357
]
},
"weight_decay": {
"feasible_points": [
0.1679001017957879
]
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
{
"learning_rate": {
"feasible_points": [
0.04037951750205473
]
},
"beta1": {
"feasible_points": [
0.9932215932637941
]
},
"beta2": {
"feasible_points": [
0.9425306939334134
]
},
"warmup_steps": {
"feasible_points": [
542
]
},
"weight_decay": {
"feasible_points": [
0.14877061239151607
]
}
}
Loading

0 comments on commit 231fc5b

Please sign in to comment.